Login| Sign Up| Help| Contact|

Patent Searching and Data


Title:
METHOD AND APPARATUS FOR COLLECTING MEDIA USAGE INFORMATION
Document Type and Number:
WIPO Patent Application WO/2019/224419
Kind Code:
A1
Abstract:
An approach is provided for collecting media usage information. The approach involves initiating, by an apparatus, a capture of at least one media sample of an ambient sound environment in proximity to the apparatus. The approach also involves comparing the at least one media sample to at least one reference media stored in the apparatus to determine a match between the at least one reference media and the at least one reference media. The approach further involves generating a data record indicating, at least in part, whether the match is determined. The approach then involves transmitting the data record from the apparatus as media usage information.

Inventors:
UR SHMUEL (IL)
Application Number:
PCT/FI2018/050386
Publication Date:
November 28, 2019
Filing Date:
May 22, 2018
Export Citation:
Click for automatic bibliography generation   Help
Assignee:
NOKIA TECHNOLOGIES OY (FI)
International Classes:
H04N21/25; G06F21/62; G06Q30/02; G10L17/08; H04W12/02
Foreign References:
US20060059277A12006-03-16
US9936249B12018-04-03
US5481294A1996-01-02
Attorney, Agent or Firm:
NOKIA TECHNOLOGIES OY et al. (FI)
Download PDF:
Claims:
CLAIMS

WHAT IS CLAIMED IS:

1. An apparatus comprising:

at least one processor; and

at least one memory including computer program code for one or more programs, the at least one memory and the computer program code configured to, with the at least one processor, cause the apparatus to perform at least the following,

receive media usage information from one or more user devices, wherein the media usage information includes one or more data records indicating whether at least one media sample captured by the one or more user devices to represent an environment in proximity to the one or more devices matches at least one reference media stored locally in the one or more user devices; and

compiling the media information into collective media usage information for the one or more user devices, the at least one reference media, or a combination thereof.

2. The apparatus according to claim 1, wherein the apparatus is further caused to perform processing the collective media usage information to generate data records identifying which of the one or more user devices has consumed the at least one reference media.

3. The apparatus according to claim 1, wherein the at least one reference media is associated with at least one time, and wherein a capture of the least one media sample is initiated at the least one time.

4. The apparatus according to claim 3, wherein the capture of the at least one media sample comprises activating a microphone of the apparatus to perform the capture for a predetermined duration at or within a threshold time window of the at least one time.

5. The apparatus according to any of claims 3 and 4, wherein a length of the at least one reference media, the threshold time window, or a combination thereof is based, at least in part, on a specificity of the at least one time.

6. The apparatus according to any of claims 1-5, wherein the at least one reference media is associated with at least one reference location.

7. The apparatus according to claim 6, wherein the apparatus is further caused to determine a location of the apparatus, and wherein the capture of the at least one media sample is based determining that the location of the apparatus is within a threshold distance of the at least one reference location. 8. The apparatus according to any of claims 1-7, wherein the apparatus is further caused to record at least one capture time of the at least one media sample, and wherein the data record further comprises a third data field indicating the at least one capture time.

9. The apparatus according to any of claims 1-8, wherein a capture of the at least one media sample is initiated by an application executing on the apparatus. 10. The apparatus according to claim 9, wherein the application is executed in an active state for a duration of the capture of the least one media sample, the comparing of the at least one media sample, the generating of the data record, the transmitting of the data record, or a combination thereof and returned to an inactive state after the duration.

11. The apparatus according to any of claims 1-10, wherein the apparatus is further caused to delete the at least one media sample, the at least one reference media, or a combination thereof after the comparing of the at least one media sample.

12. The apparatus according to any of claims 1-11, wherein the at least one reference media is downloaded to the one or more user devices over a non-cellular network connection.

13. The apparatus according to any of claims 1-12, wherein the at least one reference media is encrypted for storage at the one or more user devices.

14. The apparatus according to claim 13, wherein the media usage information is crowd-sourced. 15. A method comprising:

initiating, by an apparatus, a capture of at least one media sample of an ambient sound environment in proximity to the apparatus;

comparing the at least one media sample to at least one reference media stored in the apparatus to determine a match between the at least one reference media and the at least one reference media;

generating a data record indicating, at least in part, whether the match is determined; and

transmitting the data record from the apparatus as media usage information.

16. The method according to claim 15, wherein the at least one reference media is associated with at least one time, and wherein the capture of the least one media sample is initiated at the least one time.

17. The method according to claim 16, wherein the capture of the at least one media sample comprises activating a microphone of the apparatus to perform the capture for a predetermined duration at or within a threshold time window of the at least one time. 18. The method according to any of claims 16 and 17, wherein a length of the at least one reference media, the threshold time window, or a combination thereof is based, at least in part, on a specificity of the at least one time.

19. The method according to any of claims 15-18, wherein the at least one reference media is associated with at least one reference location.

20. The method according to claim 19, wherein the method further comprises determining a location of the apparatus, and wherein the capture of the at least one media sample is based determining that the location of the apparatus is within a threshold distance of the at least one reference location. 21. The method according to any of claims 15-20, wherein the method further comprises recording at least one capture time of the at least one media sample, and wherein the data record further comprises a third data field indicating the at least one capture time.

22. The method according to any of claims 15-21, wherein the capture of the at least one media sample is initiated by an application executing on the apparatus. 23. The method according to claim 22, wherein the application is executed in an active state for a duration of the capture of the least one media sample, the comparing of the at least one media sample, the generating of the data record, the transmitting of the data record, or a combination thereof and returned to an inactive state after the duration.

24. The method according to any of claims 15-23, wherein the method further comprises deleting the at least one media sample, the at least one reference media, or a combination thereof after the comparing of the at least one media sample.

25. The method according to any of claims 15-24, wherein the at least one reference media is downloaded to the apparatus over a non-cellular network connection.

26. The method according to any of claims 15-25, wherein the at least one reference sound is encrypted for storage at the apparatus.

27. The method according to any of claims 15-26, wherein the media usage information of the apparatus is combined with other media usage information from other apparatuses to determine collective media usage information.

28. The method according to claim 27, wherein the media usage information, the other media usage information, or a combination thereof is crowd- sourced.

29. A computer-readable storage medium carrying one or more sequences of one or more instructions which, when executed by one or more processors, cause an apparatus to perform:

initiating a capture of at least one media sample of an ambient sound environment in proximity to the apparatus;

comparing the at least one media sample to at least one reference media stored in the apparatus to determine a match between the at least one reference media and the at least one reference media;

generating a data record indicating, at least in part, whether the match is determined; and

transmitting the data record from the apparatus as media usage information.

30. The apparatus according to claim 29, wherein the at least one reference media is associated with at least one time, and wherein the capture of the least one media sample is initiated at the least one time.

31. The apparatus according to claim 30, wherein the capture of the at least one media sample comprises activating a microphone of the apparatus to perform the capture for a predetermined duration at or within a threshold time window of the at least one time. 32. The apparatus according to any of claims 30 and 31 , wherein a length of the at least one reference media, the threshold time window, or a combination thereof is based, at least in part, on a specificity of the at least one time.

33. The apparatus according to any of claims 29-32, wherein the at least one reference media is associated with at least one reference location.

34. The apparatus according to claim 33, wherein the apparatus is further caused to perform determining a location of the apparatus, and wherein the capture of the at least one media sample is based determining that the location of the apparatus is within a threshold distance of the at least one reference location. 35. The apparatus according to any of claims 29-34, wherein the apparatus is further caused to perform recording at least one capture time of the at least one media sample, and wherein the data record further comprises a third data field indicating the at least one capture time.

36. The apparatus according to any of claims 29-35, wherein the capture of the at least one media sample is initiated by an application executing on the apparatus.

37. The apparatus according to claim 36, wherein the application is executed in an active state for a duration of the capture of the least one media sample, the comparing of the at least one media sample, the generating of the data record, the transmitting of the data record, or a combination thereof and returned to an inactive state after the duration. 38. The apparatus according to any of claims 29-37, wherein the apparatus is further caused to perform deleting the at least one media sample, the at least one reference media, or a combination thereof after the comparing of the at least one media sample.

39. The apparatus according to any of claims 29-38, wherein the at least one reference media is downloaded to the apparatus over a non-cellular network connection. 40. The apparatus according to any of claims 29-39, wherein the at least one reference sound is encrypted for storage at the apparatus.

41. The apparatus according to any of claims 29-40, wherein the media usage information of the apparatus is combined with other media usage information from other apparatuses to determine collective media usage information.

42. The apparatus according to claim 41, wherein the media usage information, the other media usage information, or a combination thereof is crowd- sourced.

43. An apparatus comprising means for performing at least a method of any of claims 15-28. 44. An apparatus of claim 43, wherein the apparatus is a mobile phone further comprising:

user interface circuitry and user interface software configured to facilitate user control of at least some functions of the mobile phone through use of a display and configured to respond to user input; and

a display and display circuitry configured to display at least a portion of a user interface of the mobile phone, the display and display circuitry configured to facilitate user control of at least some functions of the mobile phone.

45. A computer program product including one or more sequences of one or more instructions which, when executed by one or more processors, cause an apparatus to at least perform the steps of at least a method of any of claims 15-28.

46. A method comprising facilitating access to at least one interface configured to allow access to at least one service, the at least one service configured to perform at least a method of any of claims 15-28.

47. A method comprising facilitating a processing of and/or processing (1) data and/or (2) information and/or (3) at least one signal, the (1) data and/or (2) information and/or (3) at least one signal based, at least in part, on at least a method of any of claims 15-28.

Description:
METHOD AND APPARATUS FOR

COLLECTING MEDIA USAGE INFORMATION

BACKGROUND

[0001] Collecting information on how and when users consume media (e.g., listen to music, watch television, hear an advertisement, etc.) presents significant technical challenges to service providers and device manufacturers (e.g., wireless, cellular, etc.). These challenges, for instance, can relate to automating the collection of media usage information from a wide user base while also protecting user privacy and minimizing resource burdens (e.g., with respect to user device resources, network resources, service provider resources, etc.).

SOME EXAMPLE EMBODIMENTS

[0002] Therefore, there is a need for an approach for efficiently collecting media usage information in a way that reduces exposure of private user information beyond a user device or apparatus (e.g., a mobile device such as a smartphone).

[0003] According to one embodiment, an apparatus comprises at least one processor, and at least one memory including computer program code for one or more computer programs, the at least one memory and the computer program code configured to, with the at least one processor, cause, at least in part, the apparatus to receive media usage information from one or more user devices. The media usage information, for instance, includes one or more data records indicating whether at least one media sample captured by the one or more user devices to represent an environment in proximity to the one or more devices matches at least one reference media stored locally in the one or more user devices. The apparatus is also caused to compile the media information into collective media usage information for the one or more user devices, the at least one reference media, or a combination thereof.

[0004] According to another embodiment, a method comprises receiving media usage information from one or more user devices. The media usage information, for instance, includes one or more data records indicating whether at least one media sample captured by the one or more user devices to represent an environment in proximity to the one or more devices matches at least one reference media stored locally in the one or more user devices. The method also comprises compiling the media information into collective media usage information for the one or more user devices, the at least one reference media, or a combination thereof

[0005] According to another embodiment, a computer-readable storage medium carrying one or more sequences of one or more instructions which, when executed by one or more processors, cause, at least in part, an apparatus to receive media usage information from one or more user devices. The media usage information, for instance, includes one or more data records indicating whether at least one media sample captured by the one or more user devices to represent an environment in proximity to the one or more devices matches at least one reference media stored locally in the one or more user devices. The apparatus is also caused to compile the media information into collective media usage information for the one or more user devices, the at least one reference media, or a combination thereof.

[0006] According to another embodiment, an apparatus comprises means for receiving media usage information from one or more user devices. The media usage information, for instance, includes one or more data records indicating whether at least one media sample captured by the one or more user devices to represent an environment in proximity to the one or more devices matches at least one reference media stored locally in the one or more user devices. The apparatus also comprises means for compiling the media information into collective media usage information for the one or more user devices, the at least one reference media, or a combination thereof.

[0007] According to another embodiment, an apparatus comprises at least one processor, and at least one memory including computer program code for one or more computer programs, the at least one memory and the computer program code configured to, with the at least one processor, cause, at least in part, the apparatus to initiate a capture of at least one media sample representing an environment in proximity to the apparatus. The apparatus is also caused to compare the at least one media sample to at least one reference media stored in the apparatus to determine a match between the at least one reference media and the at least one reference media. The apparatus is further caused to generate a data record indicating, at least in part, whether the match is determined. The apparatus is further caused to transmit the data record from the apparatus as media usage information (e.g., without exposing the captured media sample beyond the apparatus). [0008] According to one embodiment, a method comprises initiating, by an apparatus, a capture of at least one media sample of an ambient sound environment in proximity to the apparatus. The method also comprises comparing the at least one media sample to at least one reference media stored in the apparatus to determine a match between the at least one reference media and the at least one reference media. The method further comprises generating a data record indicating, at least in part, whether the match is determined. The method further comprises transmitting the data record from the apparatus as media usage information.

[0009] According to another embodiment, a computer-readable storage medium carrying one or more sequences of one or more instructions which, when executed by one or more processors, cause, at least in part, an apparatus to initiate a capture of at least one media sample representing an environment in proximity to the apparatus. The apparatus is also caused to compare the at least one media sample to at least one reference media stored in the apparatus to determine a match between the at least one reference media and the at least one reference media. The apparatus is further caused to generate a data record indicating, at least in part, whether the match is determined. The apparatus is further caused to transmit the data record from the apparatus as media usage information.

[0010] According to another embodiment, an apparatus comprises means for initiating a capture of at least one media sample of an ambient sound environment in proximity to the apparatus. The apparatus also comprises means for comparing the at least one media sample to at least one reference media stored in the apparatus to determine a match between the at least one reference media and the at least one reference media. The apparatus further comprises means for generating a data record indicating, at least in part, whether the match is determined. The apparatus further comprises means for transmitting the data record from the apparatus as media usage information.

[0011] In addition, for various example embodiments of the invention, the following is applicable: a method comprising facilitating a processing of and/or processing (1) data and/or (2) information and/or (3) at least one signal, the (1) data and/or (2) information and/or (3) at least one signal based, at least in part, on (including derived at least in part from) any one or any combination of methods (or processes) disclosed in this application as relevant to any embodiment of the invention. [0012] For various example embodiments of the invention, the following is also applicable: a method comprising facilitating access to at least one interface configured to allow access to at least one service, the at least one service configured to perform any one or any combination of network or service provider methods (or processes) disclosed in this application.

[0013] For various example embodiments of the invention, the following is also applicable: a method comprising facilitating creating and/or facilitating modifying (1) at least one device user interface element and/or (2) at least one device user interface functionality, the (1) at least one device user interface element and/or (2) at least one device user interface functionality based, at least in part, on data and/or information resulting from one or any combination of methods or processes disclosed in this application as relevant to any embodiment of the invention, and/or at least one signal resulting from one or any combination of methods (or processes) disclosed in this application as relevant to any embodiment of the invention.

[0014] For various example embodiments of the invention, the following is also applicable: a method comprising creating and/or modifying (1) at least one device user interface element and/or (2) at least one device user interface functionality, the (1) at least one device user interface element and/or (2) at least one device user interface functionality based at least in part on data and/or information resulting from one or any combination of methods (or processes) disclosed in this application as relevant to any embodiment of the invention, and/or at least one signal resulting from one or any combination of methods (or processes) disclosed in this application as relevant to any embodiment of the invention.

[0015] In various example embodiments, the methods (or processes) can be accomplished on the service provider side or on the mobile device side or in any shared way between service provider and mobile device with actions being performed on both sides.

[0016] Still other aspects, features, and advantages of the invention are readily apparent from the following detailed description, simply by illustrating a number of particular embodiments and implementations, including the best mode contemplated for carrying out the invention. The invention is also capable of other and different embodiments, and its several details can be modified in various obvious respects, all without departing from the spirit and scope of the invention. Accordingly, the drawings and description are to be regarded as illustrative in nature, and not as restrictive.

BRIEF DESCRIPTION OF THE DRAWINGS

[0017] The embodiments of the invention are illustrated by way of example, and not by way of limitation, in the figures of the accompanying drawings:

[0018] FIG. 1 is a diagram of a system capable of collecting media usage information, according to an embodiment;

[0019] FIG. 2 is a diagram of the components of a media platform or media application, according to an embodiment;

[0020] FIG. 3 is a flowchart of a process for collecting media usage information, according to one embodiment;

[0021] FIG. 4 is a diagram of a data structure for storing a media usage information, according to one embodiment;

[0022] FIGs. 5A and 5B are diagrams of user interfaces that can be used in the process for collecting media usage information, according to various embodiments;

[0023] FIG. 6 is a diagram of hardware that can be used to implement an embodiment;

[0024] FIG. 7 is a diagram of a chip set that can be used to implement an embodiment; and

[0025] FIG. 8 is a diagram of a mobile terminal (e.g., handset) that can be used to implement an embodiment.

DESCRIPTION OF SOME EMBODIMENTS

[0026] Examples of a method, apparatus, and computer program for collecting media usage information are disclosed. In the following description, for the purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the embodiments of the invention. It is apparent, however, to one skilled in the art that the embodiments of the invention may be practiced without these specific details or with an equivalent arrangement. In other instances, well-known structures and devices are shown in block diagram form in order to avoid unnecessarily obscuring the embodiments of the invention.

[0027] FIG. 1 is a diagram of a system capable of collecting media usage information, according to an embodiment. The technical field addressed by the various embodiments describe herein relate, for instance, to mobile implementation of collecting media usage information. Media usage information can include data indicating audio, video, and/or any other media data that has been consumed (e.g., heard, viewed, used, etc.) by a user. For example, there are companies that conduct surveys on what people see on the radio, television (TV), etc. including determining what specific media programs or advertisements they are exposed. This exposure can happen in non-computer related environments (e.g., through radio or TV broadcasts) or computer-related environments (e.g., consuming media through computer-based applications or services). However, tracking media exposure or consumption in non-computer related environments can be particularly challenging, often requiring manual surveys or installation of specialized tracking equipment. As a result, only a limited a number of companies [e.g., TGI (kantarmedia), Nielsen, etc.] have the resources to conduct the surveys or implement tracking equipment to track media usage in the real world.

[0028] Collecting media usage information can be big business. For example, media usage information (e.g., user exposure to advertisements or other media) can have impacts on the value of TV for advertising, on advertisers, etc. Historically, traditional media usage collection involves choosing a small number of representative people and asking them many questions related to what media they have been exposed to (e.g., viewed, heard, etc.), when the media usage occur, how the media usage occurred, etc. As noted above, traditional media usage collection can also possibly involve installing devices that check what the representative people do (e.g., with respect to media usage or consumption) and follow them around. These devices, for instance, can record what the TV does (e.g., what program is playing, what station the TV is tuned to, etc.) and who is watching.

[0029] These traditional approaches can create a few issues that can create significant technical challenges to address. First, conducting surveys and/or installing specialized equipment can be resource intensive. As a result, the sample size (e.g., the number of representative people chosen to be surveyed or tracked) are often small. This small sample size, in turn, can be too small to give a significant result particularly when tracking niche media markets or programs. For example, for some local TV stations or radio programs, a national survey may have very bad coverage for figuring out how many people are listening in a local market or to a media program that has a very small audience.

[0030] A second issue is that the representative people selected for surveys or tracking do not have privacy because they typically need to divulge everything related to their media consumption. For example, specialized tracking equipment (e.g., devices and/or applications) can be configured to always listen to what a tracked or selected person hears or views. The device then uploads the audio samples collected from this listening to a cloud or server component to compare against known media files to find a match. This way, for example, if a user is listening to music, the user’s device can sample the music that the user is listening to, send the sample to a server to identify what media is being used or consumed by the user. However, people generally can be very unhappy or wary of devices and/or applications that have such traditional listening capability because they can also potentially capture and upload privacy sensitive user information (e.g., capture user conversations or similar data) in addition to samples of the tracked media.

[0031] A third issue is that traditional media surveys and tracking is not very granular or continuous. For example, the media usage information resulting from these traditional approaches may be linked to relatively large periods of time (e.g., 1 day, 12 hours, etc.), and may not be indicative of what happened half an hour ago or even a shorter period of time (e.g., minutes ago). In other words, traditional media surveys do not produce real-time results to support, for instance, making real time decisions.

[0032] Yet another issue is that traditional survey approaches need to make sure that either the people that are used for the survey are not known, or that there are quite a lot of them. Otherwise an outside entity or actor can influence the participants, and as a result influence the result of the survey. This can then further impact how much is paid by advertisers to media providers there is a lot of incentive to influence those people. As indicated above, having a large number of participants in traditional surveys can be resource intensive and expensive, thereby leading to many traditional surveying companies keep the number of representative people small while keeping their identities secret to reduce to the potential that they survey can be unduly influenced. The need for secrecy then makes these traditional surveys more vulnerable to potential exposure of participant identities.

[0033] To address these issues and technical problems, a system 100 of FIG. 1 introduces a capability to collect media usage information 101 from a potentially large number of people using their respective mobile devices (e.g., user equipment (UE) l03a-l03n, also collectively referred to as UEs 103), while also preserving user privacy and minimizing device (e.g., UE 103) and network (e.g., a communication network 105) resources that are used for collecting the media usage information 101. More specifically, the system 100 checks if someone is listening to something at a specific time using his or her smartphone or equivalent device (e.g., UE 103).

[0034] In one embodiment, the system 100 uses a media platform 107 (e.g., a cloud component) alone or in combination with respective media applications l09a-l09n (also collectively referred to as media applications 109) executing on the UEs 103 to collect media usage information. The media applications 109 of the UEs 103, for instance, can use respective microphones 11 la-l 1 ln (also collectively referred to as microphones 111) and/or other equivalent media capture sensor (e.g., a camera sensor), to capture audio samples of media (e.g., TV or radio program, advertisement, etc.) being played by a nearby media source 113 (e.g., a TV, radio, live concert, etc.). The media application 109 then compares the captured media sample against known reference samples stored locally in the respective reference media databases H5a-l l5n (also collectively referred to as reference media databases 115) to determine a match. Data indicating whether there is a match is used to generate the media usage information 101 that can be transmitted to the media platform 107 for storage in the cloud in the media usage database 117). In other words, the captured media sample is not transmitted beyond the UE 103, and all processing of the captured media sample is performed locally at the device.

[0035] The embodiments described herein provide for several technical advantages. For example, because the media usage data collection can be performed locally at the UEs 103 via the media applications 109 (e.g., a smartphone or mobile device application), the media application 109 can benefit from distribution to a wider community of users. In this way, the system 100 can use the crowd to collect media usage data to advantageously generate a lot more media usage sampling points, and further enhance the robustness of the resulting media usage database 117. In addition, the greater participation numbers from crowd sourcing can also reduce or eliminate the need to keep the identities of participating users a secret because a larger number of participating users can reduce the potential that any single outside entity can influence the media usage habits of a large number of users.

[0036] Another advantage of the embodiments described herein for collection media usage information includes very reduced privacy issues because the media applications 109 send back very little information to the cloud component (e.g., the media platform 107). For example, in one embodiment, the media usage information 101 transmitted from the UE 103 can include just an indication of the reference sample and whether that reference sample matched an audio sample captured by the media application 109. The transmitted media usage information 101, for instance, would not include the actual sample audio or media samples collected by the microphones 111 or other equivalent sensors. Therefore, the potential exposure of any privacy sensitive information (e.g., recorded user conversations, sounds, etc.) are greatly reduced.

[0037] In one embodiment, the media application 109 need not be active continuously and need only be triggered at specific times to collected media usage information. For example, a reference media sample in the local reference media database 115 can be associated with time that the reference media is expected to be broadcast or heard by the user. A reference sample that is, for instance, an advertisement can include a date and time that the advertisement is expected to be broadcast. The media application 109 can then be activated at the expected time to monitor for the expected broadcast and deactivated after the time window has passed. As a result, the embodiments described herein can advantageously reduce the power requirement of the media application 109 by executing the media application 109 only when a reference media sample is expected to be played (e.g., as opposed to executing the media application 109 continuously in an always listening mode).

[0038] In one embodiment, power requirements of the media application 109 can be further reduced by selectively determining when the media application 109 downloads the reference media samples from a reference media database 119 of the media platform 107. For example, the media application 109 can download new reference media samples when the UE 103 is connected to an external power source (e.g., not running on batter power), when the user is not using the UE 103, when the UE 103 is connected to WiFi instead of a cellular network to reduce bandwidth costs and power requirement for transmission, and/or the like.

[0039] By way of example, as shown in FIG. 2, the media platform 107 and/or media application 109 may include one or more components for collecting media usage information according to the embodiments described herein. In one embodiment, media platform 107 and/or media application 109 include a sampling module 201 , a media comparison module 203, a usage data generator 205, and a communication module 207. It is contemplated that the functions of these components may be combined in one or more components or performed by other components with similar functionalities [e.g., a services platform 121, any of the services l23a-l23i (also collectively referred to as services 123) of the services platform 121, one or more content providers l25a-l25j (also collectively referred to as content providers 125), etc.]. The above presented modules and components of the media platform 107 and/or media application 109 can be implemented in hardware, firmware, software, or a combination thereof. Though depicted as separate entities in FIG. 1, it is contemplated that the media platform 107 and/or media application 109 may be implemented as a module of any of the components of the system 100. In another embodiment, one or more of the modules 201-207 may be implemented as a cloud-based service, local service, native application, or combination thereof. The functions of the media platform 107, media application 109, and modules 201-207 are discussed with respect to FIGs. 3-5B below.

[0040] FIG. 3 is a flowchart of a process for collecting media usage information, according to one embodiment. In one embodiment, the media platform 107, media application 109, and/or any of its modules 201-207 may perform one or more portions of the process 300 and may be implemented in, for instance, a chip set including a processor and a memory as shown in FIG. 7. As such, the media platform 107, media application 109, and/or modules 201-207 can provide means for accomplishing various parts of the process 300. In addition or alternatively, the services platform 121, services 123, and/or content providers 125 may perform any combination of the steps of the process 300 in combination with the media platform 107 or media application 109, or as standalone components. Although the process 300 is illustrated and described as a sequence of steps, it is contemplated that various embodiments of the process 300 may be performed in any order or combination and need not include all of the illustrated steps. [0041] The media platform 107 and/or media application 109 is part of technology system that can be used for collecting media usage information according to the embodiments described herein. In one embodiment, as part of this technology system, the media application 109 can be installed on respective UEs 103 (e.g., a smartphone, mobile device, computer, etc.). Accordingly, in step 301, the UE 103 can receive a query from the media application 109. In one embodiment, the query may be in the form of a reference media sample (e.g., a WAV file) of a TV program, a commcrcial/advcrtiscmcnt, and/or other media that is to be monitored for. The query can also optionally specify or otherwise associated with a time that the reference media sample is expected to be broadcast or heard by the user of the UE 103.

[0042] In one embodiment, if the expected broadcast time of reference media sample is known to a high precision (e.g., within a minutes or seconds from a scheduled time to a broadcast time like for a TV broadcast with a precisely schedule time), the reference media sample file can be a small file associated with a very specific time (e.g., specificity down to an expected second, minute, etc.). If a media source has a history of precisely scheduling a media broadcast and perform and then initiating that broadcast or performance within a precise time window of the scheduled time, the media platform 107 can generate reference media files for that include reference samples that are based on how precisely the broadcast is scheduled and/or the size of the time window between the scheduled and actual times. For example, if an advertisement is scheduled at exactly 6:00:00 pm, and the corresponding broadcast always starts within a 1 second window of the scheduled time, the media platform 107 can generate a reference sample that includes the media (e.g., sound file, video file, etc.) that would be played between 5:59:59 and 6:00:01. In contrast, if the media is a live performance or broadcast that that is scheduled for approximately 6:00:00 but can occur any time between 5:45:00 and 6: 15:00, the media platform 107 can either increase the size or length of the reference sample or keep the reference sample a short length but monitor for that shorter length reference sample over a longer time window.

[0043] In yet another embodiment, the media platform 107 can vary the length of reference samples based on their respective uniqueness (e.g., uniqueness of their sound signatures). For example, the media platform 107 may generate longer reference samples for reference media that have more similar sound signatures and generate shorter reference samples for reference media with more unique sound signatures. It is noted that broadcast time specificity and sound signature uniqueness are provided by way of illustration and not as limitations. Accordingly, it is contemplated that the media platform 107 can any characteristic of the media file, broadcast medium, performance venue, media device, UE 103, etc. to determine a length or characteristic (e.g., sampling rate, sampling format, etc.) to generate the reference media files for delivery or downloading to the media application 109.

[0044] In step 303, based on the query of step 301 or equivalent trigger, the sampling module 201 of the media application 109 initiate a capture of at least one media sample representing an environment in proximity to the apparatus (e.g., the UE 103). In one embodiment, the capture media is a sound file (e.g., a WAV file) that records the ambient sound in the environment around the UE 103 including sounds produced by any media sources near the UE 103 that may be actively outputting media. As described above, in one embodiment, the reference media is associated with at least one time (e.g., a time at which the media application 109 should be monitoring for the reference media, a time when the reference media is expected to be broadcast or performed or equivalent). In this embodiment, the sampling module 201 can initiate the capture at the associated time. For example, the capture of the at least one media sample comprises activating a microphone of the UE 103 to perform the capture for a predetermined duration at or within a threshold time window of the associated time. In one embodiment, the length of the predetermined duration or threshold time window can be based on the specificity at which the associated time is specified (e.g., specified down to the second, minute, hour, etc.). For example, if the time is specified with accuracy to the second, the sampling duration or time window can also be in seconds. Conversely if the time is specified with less specificity (e.g., specified with accuracy or precision to a minute, hour, etc.), then the duration can be minutes (e.g., if the time is accurate only to the minute), hours (e.g., if the time is accurate only to the hour), etc.

[0045] In one embodiment, the capture of the media sample can also be location based. For example, the reference media can be associated with a reference location where the reference media is expected to be heard, used, consumed, etc. Accordingly, the sampling module 201 can interact with location sensors ofthe UE 103 to determine when the UE l03’s location is within proximity of the reference location of the reference media. In other words, the sampling module 201 can determine a location of the apparatus (e.g., UE 103). The capture of the at least one media sample can then be based determining that the location of the apparatus is within a threshold distance of the at least one reference location. In this use case, UEs 103 with the media application 109 installed and are in a certain location, can listen or capture samples at a specific time. The media platform 107 can then use this information to determine that the reference media was broadcast or performed (e.g., in the case of a live performance at a concert venue) in that place at that time.

[0046] In step 305, after obtaining a media sample of the UE l03’s environment, the media comparison module 203 can compare the at least one media sample to at least one reference media stored in the apparatus (UE 103) to determine a match between the at least one reference media and the at least one reference media. For example, in a sound file use case, the media comparison module 203 uses the reference media sample (e.g., a reference WAV file specified in the query of step 301) to identify whether the captured media file (e.g., also a WAV file) matches the specified reference media sample. It is contemplated that the media comparison module 203 can use any media comparison process (e.g., process to determine a match between two sound signatures or files) known in the art to make the comparison provided the process outputs an indication of a match or no match (e.g., within matching tolerance) between the reference and captured samples. For example, if the reference media sample contains a sound signature of a program, advertisement, etc. that is expected to be performed, matching the capture media sample against the program or advertisement window is used by the media application 109 to determine whether the program or advertisement is playing and can be heard, used, or consumed by a user of the UE 103.

[0047] In step 307, the usage data generator 205 processes the output of the comparison of step 305 to generate media usage information (e.g., information indicating whether the UE 103 has capture a media sample matching a reference sample). In one embodiment, as shown in FIG. 4, the data generator 205 generates a data record 401 indicating, at least in part, whether a match (e.g., a match field 403) is determined between the captured media file and the reference media. In one embodiment, the number of data fields included in the data record 403 can vary with the specificity of the initial query or reference media. For example, if the initial query (and/or the reference media itself) that triggers the media capture specifies the exact reference media to monitor for as well as the time and/or location to monitor, the data record 401 can include just the match field 403 to indicate whether there is a match for the reference media at the specified time and/or location. In one embodiment, the match field 403 can be a binary field indicating yes or no to a match between the reference media and the captured media sample. In this way, the length of the data record 401 can be further minimized to advantageously reduce bandwidth resources. Moreover, privacy protection can be further enhanced because exposure of the data record 401 would provide no or minimal privacy sensitive information without also having knowledge of the corresponding initial query or reference media identity.

[0048] In one embodiment, the data generator 205 can include any number of additional optional data fields depending on how much information is already known to the media platform 107 (e.g., included in the initial query or associated with the reference media as metadata). For example, when the initial query is more open ended (e.g., specifies monitoring of more than one reference media at a time), the data record 403 can be generated to include, at least in part, the match field 401 in addition a data field indicating or identifying the at least one reference media (e.g., a reference media identifier field 405) against which the match was evaluated (e.g., a tuple 405). By way of example, the reference media identifier field 405 can text or numerical field containing an identifier or name of the reference media or equivalent, or equivalent. In yet another example, the data record 401 representing the media usage information can optionally include a time field 409 to indicate, for instance, a capture time of the media sample. In potential data fields (not shown in FIG. 4) can include, but are not limited to, a capture location, media type, presence of other devices, etc.

[0049] In step 309, the communication module 207 transmits the data record generated above (e.g., media usage information) from the apparatus (e.g., from UE 103 to the media platform 107) as media usage information. In one embodiment, the communication module 207 sends the tuple 405 (e.g., <name or identifier of reference media, match (yes/no)>) with or within the time filed 407. So, basically the media application 109 sends very little information (e.g., the media usage data record consisting of the tuple 405), which cannot, in anyway, be used to listen to the UE 103 (e.g., user’s phone). In this way, the user’s privacy is protected because any potential recordings of the user are not transmitted beyond the UE 103, with only the data structure of FIG. 4 subject to potential exposure. In addition, because of the tuple 405 is compact in size, it can be transmitted using means such as text messaging (e.g., SMS service) or other data efficient transmissions to reduce data transmission costs. In addition, the compact size of the tuple 405 further enables the media application 109 to monitor multiple program segments more frequently while still minimizing resource usage. In this way, media usage information can be captured for the multiple program segments at higher frequencies (e.g., every minute).

[0050] In one embodiment, the media platform 107 receives the media usage information as generated according to the embodiments described herein from one or more user devices (e.g., one or more UEs 101). For example, the media platform 107 can be operated by a media survey company to collect information on who has viewed, heard, or otherwise consumed media. In this way, the media platform 107 can compile the media usage information collected from the UEs 101 into collective media usage information. This collective media usage information can then be processed to generate data records identifying which of the UEs 101 has consumed, viewed, heard, etc. the reference media being monitored. In one embodiment, the media usage information can be crowd-sourced the UEs 101 whose users have installed the media usage collection application and/or consented to providing such information to the media platform 107. This crowd- sourcing, for instance, can advantageously increase the number of participating UEs 101 provide more complete information for media usage survey.

[0051] In one embodiment, the embodiments of the media application 109 and media platform 107 described herein provide greater efficiency and performance than traditional approaches because, in the embodiments describe herein, the media application 109 is asked if one or more specific reference media (e.g., one specific WAV file representing a program/commercial) is being played (e.g., at a specific time and/or location) as opposed to identifying capture media in general without specifying a particular reference media sample. So, the recording is compared to just a finite number of reference files (e.g., one file at a time), which can be easily done on the UE 103 (e.g., a cell phone) itself. Because this type of comparison is simple, the comparison need not be performed in the cloud, thus avoiding the need to send raw media samples from the UE 103. The difference between the embodiments described herein and traditional approaches is similar to the hard problem of object recognition in the wild "which object is this?" compared to the much easier problem of "is it object A?" Determining“which object is this?” is more open ended and requires more computations and data than the second question of“is it object A?” which is more finite and requires much less computational power. [0052] In one embodiment, the media application 109 does not need a lot of power as it works only from time to time, not continuously. In other words, the application is executed in an active state for a duration of the capture of the least one media sample, the comparing of the at least one media sample, the generating of the data record, the transmitting of the data record, or a combination thereof and returned to an inactive state after the duration. Unlike traditional applications that listens all the time and report what is being played (e.g., Shazam), the embodiments of the media application 109 described herein wake up at specific time slots and see if a specific reference media (e.g., a WAV file) is played. Even if done 100 times a day, this will be 100 seconds of operations a day, compared to 24 hours for continuous operation.

[0053] In one embodiment, the system 100 can achieve additional efficiencies. For example, the media application 109 can download the reference media files (e.g., WAV files) to the local storage of the UE 103 (e.g., local reference media database 115) when costs are lower, or bandwidth is more available (e.g., at night when next to Wi-Fi or other non-cellular network connection). In another embodiment, the media application 109 or UE 103 can delete the at least one reference media after use (e.g., after comparing to the at least one media sample). In yet another embodiment, the reference media can be encrypted for storage at the UE 103 or in the media application 109 to prevent release of the reference media from the UE 103 and ensure that the are used for comparison. In addition, signatures can be also stored locally.

[0054] As previously discussed, one advantage of the embodiments described herein is that the media platform 107 can crowd-source the media usage data to generate a lot more sample points. For example, the media platform 107 can collection media usage information from one apparatus (e.g., UE 103) and combine that media usage information with information collected from other participating UEs 103 to determine collective media usage information for a wide-base of users.

[0055] FIGs. 5A and 5B are diagrams of example user interfaces (UIs) that can be used in the process for collecting media usage information, according to various embodiments. It is noted that the example UIs of FIGs. 5 A and 5B are provided by way of illustration and not as limitation. It is contemplated that the system 100 need not provide the illustrated UIs or provide other equivalent UIs. [0056] More specifically, FIG. 5 A illustrates an example user interface (UI) 501 that can be used to provision reference media samples for collecting media usage information as described according to various embodiments described herein. As shown, the UI 501 includes a UI element 503 for selecting a reference media file to be monitored. The reference media file can be a complete program, advertisement, and/or any other media file, or a portion thereof. For example, a user can select or upload to the media platform 107 a digital file (e.g., a WAV file, video file, etc.) of the program/advertisement or clip of the program/advertisement. In one embodiment, the UI 501 also includes a UI element 505 to associate a reference time and a UI element 507 to associate a reference location with the reference media selected in the UI element 503. The UI 501 also includes a command button 509 to generate the reference media that will be stored by the media platform 107 to collect media usage information. For example, on selecting the command button 509, the media platform 107 can process the selected reference media indicated in the UI element 503 to create a version of the reference that has been trimmed to a length based on the specificity of the reference time as described above (e.g., shorter length with greater time specificity). The media platform 107h can also standardize the file format (e.g., extract an audio track of a video file in WAV format as the reference media for comparison), and then associate the reference time and reference locations as metadata or equivalent. The processed reference media can then be stored in the server-side reference media database 119 before being downloaded to the local reference media database 115 of the UE 103.

[0057] FIG. 5B illustrates an example UI 521 that can be presented on the UE 103 in response to determining a match between a specified reference media and a capture media sample. In this example, the media application 109 queries the UE 103 to initiate a comparison of the reference media to the specified reference media and determines that there is a match. In response, the media application can initiate the presentation of the UI 521 to present a message 523 (e.g.,“Match detected for reference media”) and asks the user to whether the user wants to send a corresponding media usage report (e.g., containing media usage information generated based on the match) to the media platform 107 and provides an option 525 to send the media usage information and an option 527 to not send the media usage information.

[0058] Returning to FIG. 1, in one embodiment and as previously described, the system 100 includes UEs 103, which may be utilized to execute one or more media applications 109 including, but not limited to, multimedia applications, social networking, web browsers, user interfaces (UIs) for the applications, map applications, web clients, etc. to collect media usage information and communicate with the media platform 107, other UEs 103, service platform 121, services 123, content providers 125, and/or with other components of the system 100 directly and/or via communication network 105.

[0059] The UEs 103 may be any type of mobile terminal, fixed terminal, or portable terminal including a mobile handset, station, unit, device, healthcare diagnostic and testing devices, product testing devices, multimedia computer, multimedia tablet, Internet node, communicator, desktop computer, laptop computer, notebook computer, netbook computer, tablet computer, personal communication system (PCS) device, personal navigation device, personal digital assistants (PDAs), audio/video player, digital camera/camcorder, positioning device, television receiver, loud speakers, display monitors, radio broadcast receiver, electronic book device, game device, wrist watch, or any combination thereof, including the accessories and peripherals of these devices, or any combination thereof. It is also contemplated that the UEs can support any type of interface to the user (such as “wearable” circuitry, etc.) Further, the UEs 103 may include various sensors for collecting data associated with a user, a user’s environment, and/or with a UE 103, for example, the sensors may determine and/or capture audio, video, images, atmospheric conditions, device location, user mood, ambient lighting, user physiological information, device movement speed and direction, and the like.

[0060] In one embodiment, the UE 103 includes a location module/sensor that can determine the UE 103 location (e.g., a user’s location). The UE 103 location may be determined by a triangulation system such as a GPS, assisted GPS (A-GPS), Cell of Origin, wireless local area network triangulation, or other location extrapolation technologies. Standard GPS and A-GPS systems can use the one or more satellites to pinpoint the location (e.g., longitude, latitude, and altitude) of the UE 103. A Cell of Origin system can be used to determine the cellular tower that a cellular UE 103 is synchronized with. This information provides a coarse location of the UE 103 because the cellular tower can have a unique cellular identifier (cell-ID) that can be geographically mapped. The location module/sensor may also utilize multiple technologies to detect the location of the UE 103. GPS coordinates can provide finer detail as to the location of the UE 103. In another embodiment, the UE 103 may utilize a local area network (e.g., LAN, WLAN) connection to determine the UE 103 location information, for example, from an Internet source (e.g., a service provider).

[0061] By way of example, the communication network 105 of system 100 includes one or more networks such as a data network, a wireless network, a telephony network, or any combination thereof. It is contemplated that the data network may be any local area network (LAN), metropolitan area network (MAN), wide area network (WAN), a public data network (e.g., the Internet), short range wireless network, or any other suitable packet- switched network, such as a commercially owned, proprietary packet-switched network, e.g., a proprietary cable or fiber-optic network, and the like, or any combination thereof. In addition, the wireless network may be, for example, a cellular network and may employ various technologies including enhanced data rates for global evolution (EDGE), general packet radio service (GPRS), global system for mobile communications (GSM), Internet protocol multimedia subsystem (IMS), universal mobile telecommunications system (UMTS), etc., as well as any other suitable wireless medium, e.g., worldwide interoperability for microwave access (WiMAX), Long Term Evolution (LTE) networks, code division multiple access (CDMA), wideband code division multiple access (WCDMA), wireless fidelity (WiFi), wireless LAN (WLAN), Bluetooth®, Internet Protocol (IP) data casting, satellite, mobile ad-hoc network (MANET), and the like, or any combination thereof.

[0062] In one embodiment, the content providers 125 may include and/or have access to one or more database 119a- 119h (also collectively referred to as database 119), which may store, include, and/or have access to various content items. For example, the content providers 125 may store content items (e.g., at the database 119) provided by various users, various service providers, crowd-sourced content, and the like. In various embodiments, the content providers 125 may sort, manage, store, and/or make the content items available based on various parameters, for example, geo-location information (e.g., of a submitter, of a content item, of a requestor, of a POI, etc.), sequential order, content type (e.g., audio, video, still images, etc.), date/time of content creation and/or submission, date/time of a content request, and the like. In various embodiments, the content may include media items, maps, metadata (e.g., geo-location information, content type, content creator, etc.) associated with the content items, various POIs, and the like. [0063] In one embodiment, the media platform 107 and/or media application 109 may process, analyze, detect, aggregate, and the like, reference media and captured media samples (e.g., sound recordings). In one embodiment, the media platform 107 and/or media application 109 may process metadata associated with reference media or captured media samples for determining media usage information. Examples of the metadata include, but are not limited to, media source, user information, user preferences, user device capabilities, audio characteristics, and the like. In various embodiments, the media platform 107 and/or media application 109 may include media editing functions such as cut, copy, paste, delete, insert, silence, auto-trim, and the like. Further, audio effects including amplify, normalize, equalizer, envelope, reverb, echo, reverse, etc. can be applied. Furthermore, capabilities include support for media file formats including mp3, wav, wma, au, aac, m4a, mid, and the like.

[0064] In certain embodiments, the media platform 107 is implemented as a collection of one or more hardware, software, algorithms, firmware, or combinations thereof that can be integrated for use with the services platform 121 , services 123, and/or content providers 125. In various embodiments, the media platform 107 can be maintained on a network server, while operating in connection with the media application 109 as an extensible feature, a web-service, an applet, a script, an object-oriented application, or the like to enable processing, analyzing, detecting, and the like, of media usage information. Further, the media platform 107 and/or media applications 109 may utilize one or more service application programming interfaces (APIs)/integrated interface, through which communication, media, content, and information (e.g., associated with media usage information) may be shared, accessed and/or processed.

[0065] By way of example, the media platform 107, media application 109, and/or other components of the system 100 may communicate with each other using well known, new or still developing protocols. In this context, a protocol includes a set of rules defining how the network nodes within the communication network 105 interact with each other based on information sent over the communication links. The protocols are effective at different layers of operation within each node, from generating and receiving physical signals of various types, to selecting a link for transferring those signals, to the format of information indicated by those signals, to identifying which software application executing on a computer system sends or receives the information. The conceptually different layers of protocols for exchanging information over a network are described in the Open Systems Interconnection (OSI) Reference Model.

[0066] Communications between the network nodes are typically effected by exchanging discrete packets of data. Each packet typically comprises (1) header information associated with a particular protocol, and (2) payload information that follows the header information and contains information that may be processed independently of that particular protocol. In some protocols, the packet includes (3) trailer information following the payload and indicating the end of the payload information. The header includes information such as the source of the packet, its destination, the length of the payload, and other properties used by the protocol. Often, the data in the payload for the particular protocol includes a header and payload for a different protocol associated with a different, higher layer of the OSI Reference Model. The header for a particular protocol typically indicates a type for the next protocol contained in its payload. The higher layer protocol is said to be encapsulated in the lower layer protocol. The headers included in a packet traversing multiple heterogeneous networks, such as the Internet, typically include a physical (layer 1) header, a data- link (layer 2) header, an internetwork (layer 3) header and a transport (layer 4) header, and various application (layer 5, layer 6 and layer 7) headers as defined by the OSI Reference Model.

[0067] In one embodiment, the media platform 107 and media application 109 may interact according to a client-server model. It is noted that the client-server model of computer process interaction is widely known and used. According to the client-server model, a client process sends a message including a request to a server process, and the server process responds by providing a service. The server process may also return a message with a response to the client process. Often the client process and server process execute on different computer devices, called hosts, and communicate via a network using one or more protocols for network communications. The term“server” is conventionally used to refer to the process that provides the service, or the host computer on which the process operates. Similarly, the term“client” is conventionally used to refer to the process that makes the request, or the host computer on which the process operates. As used herein, the terms“client” and“server” refer to the processes, rather than the host computers, unless otherwise clear from the context. In addition, the process performed by a server can be broken up to run as multiple processes on multiple hosts (sometimes called tiers) for reasons that include reliability, scalability, and redundancy, among others. It is also noted that the role of a client and a server is not fixed; in some situations, a device may act both as a client and a server, which may be done simultaneously, and/or the device may alternate between these roles.

[0068] The processes described herein for collecting media usage information may be advantageously implemented via software, hardware (e.g., general processor, Digital Signal Processing (DSP) chip, an Application Specific Integrated Circuit (ASIC), Field Programmable Gate Arrays (FPGAs), etc.), firmware or a combination thereof. Such exemplary hardware for performing the described functions is detailed below.

[0069] FIG. 6 illustrates a computer system 600 upon which an embodiment of the invention may be implemented. Computer system 600 is programmed (e.g., via computer program code or instructions) to collect media usage information as described herein and includes a communication mechanism such as a bus 610 for passing information between other internal and external components of the computer system 600. Information (also called data) is represented as a physical expression of a measurable phenomenon, typically electric voltages, but including, in other embodiments, such phenomena as magnetic, electromagnetic, pressure, chemical, biological, molecular, atomic, sub-atomic and quantum interactions. For example, north and south magnetic fields, or a zero and non-zero electric voltage, represent two states (0, 1) of a binary digit (bit). Other phenomena can represent digits of a higher base. A superposition of multiple simultaneous quantum states before measurement represents a quantum bit (qubit). A sequence of one or more digits constitutes digital data that is used to represent a number or code for a character. In some embodiments, information called analog data is represented by a near continuum of measurable values within a particular range.

[0070] A bus 610 includes one or more parallel conductors of information so that information is transferred quickly among devices coupled to the bus 610. One or more processors 602 for processing information are coupled with the bus 610.

[0071] A processor 602 performs a set of operations on information as specified by computer program code related to collecting media usage information. The computer program code is a set of instructions or statements providing instructions for the operation of the processor and/or the computer system to perform specified functions. The code, for example, may be written in a computer programming language that is compiled into a native instruction set of the processor. The code may also be written directly using the native instruction set (e.g., machine language). The set of operations include bringing information in from the bus 610 and placing information on the bus 610. The set of operations also typically include comparing two or more units of information, shifting positions of units of information, and combining two or more units of information, such as by addition or multiplication or logical operations like OR, exclusive OR (XOR), and AND. Each operation of the set of operations that can be performed by the processor is represented to the processor by information called instructions, such as an operation code of one or more digits. A sequence of operations to be executed by the processor 602, such as a sequence of operation codes, constitute processor instructions, also called computer system instructions or, simply, computer instructions. Processors may be implemented as mechanical, electrical, magnetic, optical, chemical or quantum components, among others, alone or in combination.

[0072] Computer system 600 also includes a memory 604 coupled to bus 610. The memory 604, such as a random access memory (RAM) or other dynamic storage device, stores information including processor instructions for collecting media usage information. Dynamic memory allows information stored therein to be changed by the computer system 600. RAM allows a unit of information stored at a location called a memory address to be stored and retrieved independently of information at neighboring addresses. The memory 604 is also used by the processor 602 to store temporary values during execution of processor instructions. The computer system 600 also includes a read only memory (ROM) 606 or other static storage device coupled to the bus 610 for storing static information, including instructions, that is not changed by the computer system 600. Some memory is composed of volatile storage that loses the information stored thereon when power is lost. Also coupled to bus 610 is a non-volatile (persistent) storage device 608, such as a magnetic disk, optical disk or flash card, for storing information, including instructions, that persists even when the computer system 600 is turned off or otherwise loses power.

[0073] Information, including instructions for collecting media usage information, is provided to the bus 610 for use by the processor from an external input device 612, such as a keyboard containing alphanumeric keys operated by a human user, or a sensor. A sensor detects conditions in its vicinity and transforms those detections into physical expression compatible with the measurable phenomenon used to represent information in computer system 600. Other external devices coupled to bus 610, used primarily for interacting with humans, include a display device 614, such as a cathode ray tube (CRT) or a liquid crystal display (LCD), or plasma screen or printer for presenting text or images, and a pointing device 616, such as a mouse or a trackball or cursor direction keys, or motion sensor, for controlling a position of a small cursor image presented on the display 614 and issuing commands associated with graphical elements presented on the display 614. In some embodiments, for example, in embodiments in which the computer system 600 performs all functions automatically without human input, one or more of external input device 612, display device 614 and pointing device 616 is omitted.

[0074] In the illustrated embodiment, special purpose hardware, such as an application specific integrated circuit (ASIC) 620, is coupled to bus 610. The special purpose hardware is configured to perform operations not performed by processor 602 quickly enough for special purposes. Examples of application specific ICs include graphics accelerator cards for generating images for display 614, cryptographic boards for encrypting and decrypting messages sent over a network, speech recognition, and interfaces to special external devices, such as robotic arms and medical scanning equipment that repeatedly perform some complex sequence of operations that are more efficiently implemented in hardware.

[0075] Computer system 600 also includes one or more instances of a communications interface 670 coupled to bus 610. Communication interface 670 provides a one-way or two-way communication coupling to a variety of external devices that operate with their own processors, such as printers, scanners and external disks. In general, the coupling is with a network link 678 that is connected to a local network 680 to which a variety of external devices with their own processors are connected. For example, communication interface 670 may be a parallel port or a serial port or a universal serial bus (USB) port on a personal computer. In some embodiments, communications interface 670 is an integrated services digital network (ISDN) card or a digital subscriber line (DSL) card or a telephone modem that provides an information communication connection to a corresponding type of telephone line. In some embodiments, a communication interface 670 is a cable modem that converts signals on bus 610 into signals for a communication connection over a coaxial cable or into optical signals for a communication connection over a fiber optic cable. As another example, communications interface 670 may be a local area network (LAN) card to provide a data communication connection to a compatible LAN, such as Ethernet. Wireless links may also be implemented. For wireless links, the communications interface 670 sends or receives or both sends and receives electrical, acoustic or electromagnetic signals, including infrared and optical signals, that carry information streams, such as digital data. For example, in wireless handheld devices, such as mobile telephones like cell phones, the communications interface 670 includes a radio band electromagnetic transmitter and receiver called a radio transceiver. In certain embodiments, the communications interface 670 enables connection to the communication network 105 for collecting media usage information.

[0076] The term computer-readable medium is used herein to refer to any medium that participates in providing information to processor 602, including instructions for execution. Such a medium may take many forms, including, but not limited to, non-volatile media, volatile media and transmission media. Non-volatile media include, for example, optical or magnetic disks, such as storage device 608. Volatile media include, for example, dynamic memory 604. Transmission media include, for example, coaxial cables, copper wire, fiber optic cables, and carrier waves that travel through space without wires or cables, such as acoustic waves and electromagnetic waves, including radio, optical and infrared waves. Signals include man-made transient variations in amplitude, frequency, phase, polarization or other physical properties transmitted through the transmission media. Common forms of computer-readable media include, for example, a floppy disk, a flexible disk, hard disk, magnetic tape, any other magnetic medium, a CD-ROM, CDRW, DVD, any other optical medium, punch cards, paper tape, optical mark sheets, any other physical medium with patterns of holes or other optically recognizable indicia, a RAM, a PROM, an EPROM, a FFASH-EPROM, any other memory chip or cartridge, a carrier wave, or any other medium from which a computer can read.

[0077] FIG. 7 illustrates a chip set 700 upon which an embodiment of the invention may be implemented. Chip set 700 is programmed to collect media usage information as described herein and includes, for instance, the processor and memory components described with respect to FIG. 6 incorporated in one or more physical packages (e.g., chips). By way of example, a physical package includes an arrangement of one or more materials, components, and/or wires on a structural assembly (e.g., a baseboard) to provide one or more characteristics such as physical strength, conservation of size, and/or limitation of electrical interaction. It is contemplated that in certain embodiments the chip set can be implemented in a single chip.

[0078] In one embodiment, the chip set 700 includes a communication mechanism such as a bus 701 for passing information among the components of the chip set 700. A processor 703 has connectivity to the bus 701 to execute instructions and process information stored in, for example, a memory 705. The processor 703 may include one or more processing cores with each core configured to perform independently. A multi-core processor enables multiprocessing within a single physical package. Examples of a multi core processor include two, four, eight, or greater numbers of processing cores. Alternatively or in addition, the processor 703 may include one or more microprocessors configured in tandem via the bus 701 to enable independent execution of instructions, pipelining, and multithreading. The processor 703 may also be accompanied with one or more specialized components to perform certain processing functions and tasks such as one or more digital signal processors (DSP) 707, or one or more application-specific integrated circuits (ASIC) 709. A DSP 707 typically is configured to process real-world signals (e.g., sound) in real time independently of the processor 703. Similarly, an ASIC 709 can be configured to performed specialized functions not easily performed by a general purposed processor. Other specialized components to aid in performing the inventive functions described herein include one or more field programmable gate arrays (FPGA) (not shown), one or more controllers (not shown), or one or more other special-purpose computer chips.

[0079] The processor 703 and accompanying components have connectivity to the memory 705 via the bus 701. The memory 705 includes both dynamic memory (e.g., RAM, magnetic disk, writable optical disk, etc.) and static memory (e.g., ROM, CD-ROM, etc.) for storing executable instructions that when executed perform the inventive steps described herein to collect media usage information. The memory 705 also stores the data associated with or generated by the execution of the inventive steps.

[0080] FIG. 8 is a diagram of exemplary components of a mobile station (e.g., handset) capable of operating in the system of FIG. 1, according to one embodiment. Generally, a radio receiver is often defined in terms of front-end and back-end characteristics. The front-end of the receiver encompasses all of the Radio Frequency (RF) circuitry whereas the back-end encompasses all of the base-band processing circuitry. Pertinent internal components of the telephone include a Main Control Unit (MCU) 803, a Digital Signal Processor (DSP) 805, and a receiver/transmitter unit including a microphone gain control unit and a speaker gain control unit. A main display unit 807 provides a display to the user in support of various applications and mobile station functions that offer automatic contact matching. An audio function circuitry 809 includes a microphone 811 and microphone amplifier that amplifies the speech signal output from the microphone 811. The amplified speech signal output from the microphone 811 is fed to a coder/decoder (CODEC) 813.

[0081] A radio section 815 amplifies power and converts frequency in order to communicate with a base station, which is included in a mobile communication system, via antenna 817. The power amplifier (PA) 819 and the transmitter/modulation circuitry are operationally responsive to the MCU 803, with an output from the PA 819 coupled to the dup lexer 821 or circulator or antenna switch, as known in the art. The PA 819 also couples to a battery interface and power control unit 820.

[0082] In use, a user of mobile station 801 speaks into the microphone 811 and his or her voice along with any detected background noise is converted into an analog voltage. The analog voltage is then converted into a digital signal through the Analog to Digital Converter (ADC) 823. The control unit 803 routes the digital signal into the DSP 805 for processing therein, such as speech encoding, channel encoding, encrypting, and interleaving. In one embodiment, the processed voice signals are encoded, by units not separately shown, using a cellular transmission protocol such as global evolution (EDGE), general packet radio service (GPRS), global system for mobile communications (GSM), Internet protocol multimedia subsystem (IMS), universal mobile telecommunications system (UMTS), etc., as well as any other suitable wireless medium, e.g., microwave access (WiMAX), Long Term Evolution (LTE) networks, code division multiple access (CDMA), wireless fidelity (WiFi), satellite, and the like.

[0083] The encoded signals are then routed to an equalizer 825 for compensation of any frequency-dependent impairments that occur during transmission though the air such as phase and amplitude distortion. After equalizing the bit stream, the modulator 827 combines the signal with a RF signal generated in the RF interface 829. The modulator 827 generates a sine wave by way of frequency or phase modulation. In order to prepare the signal for transmission, an up-converter 831 combines the sine wave output from the modulator 827 with another sine wave generated by a synthesizer 833 to achieve the desired frequency of transmission. The signal is then sent through a PA 819 to increase the signal to an appropriate power level. In practical systems, the PA 819 acts as a variable gain amplifier whose gain is controlled by the DSP 805 from information received from a network base station. The signal is then filtered within the dup lexer 821 and optionally sent to an antenna coupler 835 to match impedances to provide maximum power transfer. Finally, the signal is transmitted via antenna 817 to a local base station. An automatic gain control (AGC) can be supplied to control the gain of the final stages of the receiver. The signals may be forwarded from there to a remote telephone which may be another cellular telephone, other mobile phone or a land- line connected to a Public Switched Telephone Network (PSTN), or other telephony networks.

[0084] Voice signals transmitted to the mobile station 801 are received via antenna 817 and immediately amplified by a low noise amplifier (LNA) 837. A down-converter 839 lowers the carrier frequency while the demodulator 841 strips away the RF leaving only a digital bit stream. The signal then goes through the equalizer 825 and is processed by the DSP 805. A Digital to Analog Converter (DAC) 843 converts the signal and the resulting output is transmitted to the user through the speaker 845, all under control of a Main Control Unit (MCU) 803-which can be implemented as a Central Processing Unit (CPU) (not shown).

[0085] The MCU 803 receives various signals including input signals from the keyboard

847. The keyboard 847 and/or the MCU 803 in combination with other user input components (e.g., the microphone 811) comprise a user interface circuitry for managing user input. The MCU 803 runs a user interface software to facilitate user control of at least some functions of the mobile station 801 to collect media usage information. The MCU 803 also delivers a display command and a switch command to the display 807 and to the speech output switching controller, respectively. Further, the MCU 803 exchanges information with the DSP 805 and can access an optionally incorporated SIM card 849 and a memory 851. In addition, the MCU 803 executes various control functions required of the station. The DSP 805 may, depending upon the implementation, perform any of a variety of conventional digital processing functions on the voice signals. Additionally, DSP 805 determines the background noise level of the local environment from the signals detected by microphone 811 and sets the gain of microphone 811 to a level selected to compensate for the natural tendency of the user of the mobile station 801.

[0086] The CODEC 813 includes the ADC 823 and DAC 843. The memory 851 stores various data including call incoming tone data and is capable of storing other data including music data received via, e.g., the global Internet. The software module could reside in RAM memory, flash memory, registers, or any other form of writable computer-readable storage medium known in the art including non-transitory computer-readable storage medium. For example, the memory device 851 may be, but not limited to, a single memory, CD, DVD, ROM, RAM, EEPROM, optical storage, or any other non-volatile or non- transitory storage medium capable of storing digital data.

[0087] An optionally incorporated SIM card 849 carries, for instance, important information, such as the cellular phone number, the carrier supplying service, subscription details, and security information. The SIM card 849 serves primarily to identify the mobile station 801 on a radio network. The card 849 also contains a memory for storing a personal telephone number registry, text messages, and user specific mobile station settings.

[0088] While the invention has been described in connection with a number of embodiments and implementations, the invention is not so limited but covers various obvious modifications and equivalent arrangements, which fall within the purview of the appended claims. Although features of the invention are expressed in certain combinations among the claims, it is contemplated that these features can be arranged in any combination and order.