CROSS-REFERENCE TO RELATED APPLICATIONThis application is a continuation of, and claims the benefit and priority of, U.S. patent application Ser. No. 15/855,275 filed on Dec. 27, 2017. The entire contents of such application are hereby incorporated herein by reference.
BACKGROUNDIt is popular to use mobile devices, such as smartphones, to record videos of various events. For example, people use smartphones to record family trips and activities, sports games, ceremonies, and performances of family members, friends and others in the fields of athletics, education, entertainment and business. Many of these events involve interesting moments that occur over long stretches of time. During the events, it can be difficult to anticipate or predict when these interesting moments will occur. Consequently, even though a viewer may wish to only capture the interesting moments, the viewer must record the entire event to avoid missing the interesting moments. To develop highlight videos, the viewers must edit these videos after the recording, which can be painstaking, time consuming and labor intensive.
Also, while recording the video, it can be difficult to take note of important information. Conventionally, this requires the use of at least two separate tools—the smartphone's video recorder and a separate software program or paper. The viewer operates the video recorder to record the event. Another person, such as a friend or statistician, uses the software program or paper to note the important information regarding the interesting moments.
For example, the statistician might note that a specific participant scored a point or made a particular action.
It can be challenging for two people to manage these separate tools especially in high-paced events. If there is only one person available to view an event, the person may decide not to use one of the tools, losing the opportunity to gain valuable video or event information. Alternatively, the person may attempt to manage both of these tools at the same time. This can cause difficulty, stress, errors and oversights in the video recording process and note-taking process.
Furthermore, there are several shortcomings in the known processes for recording, storing, publishing, finding, rating and acting upon videos of participants in events. The shortcomings include, but are not limited to, the burdens of labor and time required to edit videos after they are recorded, inefficiencies in the processes of the human machine interface, the difficulty to find videos of a desired category, the overuse of data storage centers, the loss of data storage capacity on mobile devices such as smartphones, and the inaccuracies in the event information that is published in connection with videos. These shortcomings result in disadvantages and lost opportunities for viewers who record videos, the event participants and the viewers who watch videos.
The foregoing background describes some, but not necessarily all, of the problems, disadvantages and challenges related to video recording, video management, video access, video-related activities, event reporting, and the pursuits of event participants and viewers.
BRIEF DESCRIPTION OF THE DRAWINGSFIG. 1 is a schematic, block diagram illustrating an embodiment of the system operatively coupled to devices and data sources over a network.
FIG. 2A is a top view of an embodiment of the login interface of the programmed device.
FIG. 2B is a top view of an embodiment of the user profile interface of the programmed device.
FIG. 3A is a top view of an embodiment of the home interface of the programmed device.
FIG. 3B is a top view of an embodiment of the main features interface of the programmed device.
FIG. 3C is a top view of an embodiment of the update filter interface of the programmed device.
FIG. 4 is a top view of an embodiment of the filter strips of the programmed device.
FIG. 5A is a top view of an embodiment of the map search interface of the programmed device.
FIG. 5B is a top view of an example of the map search interface ofFIG. 5A.
FIG. 6A is a top view of an embodiment of the recording options interface of the programmed device.
FIG. 6B is a top view of an embodiment of the recording features interface of the programmed device.
FIG. 7 is table illustrating an embodiment of the basic mode for recording with the programmed device.
FIG. 8 is a top view of an embodiment of the programmed device, illustrating the user's thumb touching the start/stop element to start the basic mode recording session.
FIG. 9 is a top view of an embodiment of the programmed device, illustrating the user's single finger touching the screen of the programmed device during the basic mode recording session to generate a clip input.
FIG. 10A is a top view of an embodiment of the programmed device, illustrating the flash in response to the user's clip input (e.g., touching of the screen of the programmed device) during the basic mode recording session.
FIG. 10B is a top view of an embodiment of the programmed device, illustrating the disappearance of the flash ofFIG. 10A during the basic mode recording session.
FIG. 11 is a rear view of an embodiment of the programmed device, illustrating the rear lens.
FIG. 12 is a rear view of an embodiment of the programmed device, illustrating the rear lens covered by the user's hand to end or exit the basic mode recording session.
FIG. 13A is a top view of an embodiment of the publish decision interface of the programmed device.
FIG. 13B is a top view of an embodiment of the programmed device, illustrating the programmed device oriented in a vertical or portrait position during the basic mode recording session.
FIG. 14 is table illustrating an embodiment of the advanced mode for recording video and statistics with the programmed device.
FIG. 15 is table illustrating an embodiment of the correlations for the advanced mode ofFIG. 14.
FIG. 16 is a top view of an embodiment of the programmed device, illustrating the user's single finger touching the screen to generate a clip input and record one point during the advanced mode recording session.
FIG. 17 is a top view of an embodiment of the programmed device, illustrating two fingers touching the screen to generate a clip input and record two points during the advanced mode recording session.
FIG. 18 is a top view of an embodiment of the programmed device, illustrating three fingers touching the screen to generate a clip input and record three points during the advanced mode recording session.
FIG. 19 is a top view of an embodiment of the programmed device, illustrating one finger swiping laterally on the screen to generate a clip input and record an assist during the advanced mode recording session.
FIG. 20 is a top view of an embodiment of the programmed device, illustrating one finger swiping vertically on the screen to generate a clip input and record a rebound during the advanced mode recording session.
FIG. 21 is a top view of an embodiment of the programmed device, illustrating four fingers touching the screen to generate a clip input and record a steal during the advanced mode recording session.
FIG. 22 is a top view of an embodiment of the programmed device, illustrating the base of a first or hand touching the screen to generate a clip input and record a block during the advanced mode recording session.
FIG. 23 is a top view of an embodiment of the programmed device, illustrating a finger marking an X on the screen to generate a clip input and record a turnover during the advanced mode recording session.
FIG. 24A is a top view of an embodiment of the programmed device, illustrating a recording interface having different categories of clip elements (e.g., highlight clip elements and lowlight clip elements) for the advanced mode recording session.
FIG. 24B is a top view of an embodiment of the programmed device, illustrating the recording interface ofFIG. 24A after one second has elapsed.
FIG. 25A is a top view of an embodiment of the programmed device, illustrating the recording interface ofFIG. 24A after three seconds have elapsed.
FIG. 25B is a top view of an embodiment of the programmed device, illustrating the recording interface ofFIG. 24A when the user selected a highlight clip element at the point of one minute and nineteen seconds.
FIG. 26 is a top view of an embodiment of the programmed device, illustrating the recording interface having different categories of clip elements (e.g., highlight clip elements and lowlight clip elements) and selectable statistics symbols for the advanced mode recording session.
FIG. 27 is a top view of an embodiment of a cutback pop-up of the programmed device.
FIG. 28 is the first part of a table illustrating an example of an embodiment of a data list generated by the video generator of the programmed device during a recording session.
FIG. 29 is the second part of the table ofFIG. 28.
FIG. 30A is a schematic diagram illustrating a video track generated during a period of time during a recording session of the programmed device.
FIG. 30B is a schematic diagram illustrating the bookmarking process corresponding to the data list ofFIGS. 28-29 to determine or identify excess tracks and desired clips.
FIG. 31 is the first part of a table illustrating another example of an embodiment of the data list generated by the video generator of the programmed device during a recording session.
FIG. 32 is the second part of the table ofFIG. 31.
FIG. 33 is a schematic diagram illustrating the bookmarking process corresponding to a data list ofFIGS. 31-32 to determine or identify excess tracks and desired clips.
FIG. 34 is the first part of a table illustrating yet another example of an embodiment of a data list generated by the video generator of the programmed device during a recording session.
FIG. 35 is the second part of the table ofFIG. 34.
FIG. 36 is a schematic diagram illustrating the bookmarking process corresponding to the data list ofFIGS. 34-35 to determine or identify excess tracks and desired clips.
FIG. 37 is a flow chart illustrating an embodiment of the recording method of the programmed device.
FIG. 38 is a schematic diagram illustrating the results of the recording method ofFIG. 37.
FIG. 39 is a top view of an embodiment of the processing interfaces of the programmed device.
FIG. 40A is a top view of an embodiment of the primary video categorizer interface of the programmed device.
FIG. 40B is a top view of an embodiment of the secondary video categorizer interface of the programmed device.
FIG. 40C is a top view of an embodiment of the public publication interface of the programmed device.
FIG. 41 is a top view of an embodiment of the front video interface of the programmed device.
FIG. 42A is a top view of an embodiment of the social interface of the programmed device.
FIG. 42B is a top view of an embodiment of the rating interface of the programmed device.
FIG. 43A is a top view of an embodiment of the secondary video categorizer interface ofFIG. 40B, illustrating a selection of the athlete lowlights category.
FIG. 43B is a top view of an embodiment of the private posting interface of the programmed device.
FIG. 44 is a flow chart of an embodiment of a method for verifying or confirming the accuracy of event information reported by users of programmed devices.
FIG. 45 is a flow chart of an embodiment of another method for verifying or confirming the accuracy of event information reported by users of programmed devices.
FIG. 46 is a top view of an embodiment of an outcome indicator of an event site or facility.
FIG. 47A is a top view of an embodiment of the image capture interface of the programmed device, illustrating a photo of the outcome indicator ofFIG. 46.
FIG. 47A is a top view of an embodiment of the image capture interface of the programmed device, illustrating a scoreboard photo.
FIG. 47B is a top view of an embodiment of the image capture interface of the programmed device, illustrating a photo of a physical display medium, such as a mascot banner.
FIG. 48A is a top view of an embodiment of a process indicator of the programmed device.
FIG. 48B is a top view of an embodiment of the verification success indicator of the programmed device.
FIG. 48C is a top view of an embodiment of the verification failure indicator of the programmed device.
FIG. 49A is a top view of an embodiment of the winner benefit interface of the programmed device.
FIG. 49B is a top view of an embodiment of the loser benefit interface of the programmed device.
FIG. 50A is a top view of an embodiment of the participant center interface of the programmed device.
FIG. 50B is a top view of an embodiment of the personal data interface of the programmed device.
FIG. 51A is a top view of an embodiment of the personal data verification interface of the programmed device.
FIG. 51B is a top view of an embodiment of the verification progress interface of the programmed device.
FIG. 52A is a top view of an embodiment of the highlight video interface of the programmed device.
FIG. 52B is a top view of an embodiment of the interview video interface of the programmed device.
FIG. 53A is a top view of an embodiment of the reference video interface of the programmed device.
FIG. 53B is a top view of an embodiment of the biography interface of the programmed device.
FIG. 54A is a top view of an embodiment of the send videos interface of the programmed device.
FIG. 54B is a top view of an embodiment of the recipient interface of the programmed device.
FIG. 55A is a top view of an embodiment of the lowlight video interface of the programmed device.
FIG. 55B is a top view of an embodiment of the development video interface of the programmed device.
FIG. 56 is a top view of an embodiment of the gift card interface of the programmed device.
FIG. 57A is a top view of an embodiment of the sponsor level interface of the programmed device.
FIG. 57B is a top view of an embodiment of the sponsors interface of the programmed device.
FIG. 57C is a top view of an embodiment of the sponsor account interface of the programmed device.
FIG. 58A is a top view of an embodiment of the connector interface of the programmed device.
FIG. 58B is a top view of an embodiment of the listing interface of the programmed device.
FIG. 59A is a top view of an embodiment of the connection search interface of the programmed device.
FIG. 59B is a top view of an embodiment of the search results interface of the programmed device.
FIG. 60A is a top view of an embodiment of the provider interface of the programmed device, illustrating the masking of the videos and text of the reviews.
FIG. 60B is a top view of an embodiment of the review unlock interface of the programmed device.
FIG. 61A is a top view of an embodiment of the provider interface ofFIG. 60A, illustrating the unmasked videos and text of the reviews.
FIG. 61B is a top view of an embodiment of the provider profile of the programmed device.
FIG. 62A is a top view of an embodiment of the order interface of the programmed device, illustrating an example of an order for a bracelet.
FIG. 62B is an isometric view of an embodiment of a bracelet configured to be operatively coupled to the programmed device.
FIG. 63A is a top view of an embodiment of another order interface of the programmed device, illustrating an example of an order for a shoestring tag.
FIG. 62B is a top view of an embodiment of a shoestring tag configured to be operatively coupled to the programmed device.
FIG. 63C is a schematic side view of the shoestring tag ofFIG. 62B.
FIG. 64A is a top view of the shoestring tag ofFIG. 62B, illustrating the coupling of the shoestring tag to a shoestring.
FIG. 64B is an isometric view of an embodiment of a shoe having the shoestring tag ofFIG. 62B.
FIG. 65 is a top view of an embodiment of the athlete metrics interface of the programmed device.
FIG. 66 is a top view of an embodiment of certain video footage (e.g., the dribbling player's feet) tracked by the tracking images generated by the programmed device.
FIG. 67 is a table illustrating an embodiment of an animation set generated by the programmed device.
DETAILED DESCRIPTIONAs illustrated inFIG. 1, in an embodiment, thesystem10 is stored within one or more databases ordata storage devices12. The one or moredata storage devices12 are accessible to one or more processors, such asprocessor14, over adata network16, such as the Internet. Theprocessor14 is operatively coupled to a plurality ofdata sources18 over thedata network16. Users can operate a plurality of types ofelectronic devices20 to access thesystem10 through thenetwork16. Theelectronic devices20 can include apersonal computer22,smartphone24,tablet26 or any other type of network access device.
Thesystem10 includes a plurality of computer-readable instructions, software, computer code, computer programs, logic, algorithms, data, data libraries, data files, graphical data and commands that are executable by theprocessor14 and theelectronic devices20. In operation, theprocessor14 and theelectronic devices20 cooperate with thesystem10 to perform the functions described in this description.
In an embodiment, thesystem10 includes avideo generator28,interface module30,publication module31,participant module32,verification module34 andconnector module36. The one or moredata storage devices12 store thesystem10 for execution by theprocessor14. Theelectronic devices20 can access thesystem10 over thenetwork16 to enable users to provide inputs and receive outputs as described below.
In addition, the one or moredata storage devices12 store adownloadable system11. In an embodiment, thedownloadable system11 includes part or all of thesystem10 in a format that is configured to be downloaded and installed onto theelectronic devices20. For example, in an embodiment, thedownloadable system11 includes: (a) a mobile app version of thesystem10 that is compatible with the iOS™ mobile operating system; and (b) a mobile app version of thesystem10 that is compatible with the Android™ mobile operating system. In an embodiment, thedata sources18 include databases of schools38, databases ofhealthcare providers40, databases oftesting organizations42, databases ofbenefit sources44 and databases ofsponsors46.
From time to time in this description, thesystem13, which includes thesystems10 and11 or portions thereof, may be described as performing various functions with the understanding that such functions involve the execution role of theprocessor14, another processor or theelectronic devices20. Depending upon the embodiment, theprocessor14 and theelectronic devices20 can include one or more microprocessors, circuits, circuitry, controllers or other data processing devices. Although thesystem13 is operable to control the input and output devices of theelectronic devices20, thesystem13 may be described herein as generating outputs, displaying interfaces and receiving inputs.
Theelectronic devices20 are configured to download, store and execute thedownloadable system11. As illustrated inFIG. 2, once downloaded on one of theelectronic devices20, thedownloadable system11 causes theelectronic device20 to perform various functions. The term, programmeddevice120, may be used herein to refer to anelectronic device20 that is operable according to, or based on the commands, instructions and functionality of thesystem13, including thedownloadable system11.
There are a variety of different types of users of the programmeddevices120 and thesystem13, including, but not limited to, event participants (e.g., students and athletes), family members and friends of event participants, news media professionals and journalists, video producers, schools, colleges, coaches, sponsors of event participants, merchants (e.g., restaurants) and providers (e.g., sports clubs/teams, camp hosts, college recruiters, physical therapists, sports agents, trainer, academic tutors and others).
In an embodiment, the programmeddevice120 includes an imaging device configured to record videos and generate images or photographs. The imaging device can include dual cameras or a camera unit with dual lenses (one for front imaging and one for rear imaging) to detect the user's gestures at the front while recording videos of action at the rear. In an embodiment, the imaging device has auto-zoom (zoom-in and zoom-out) functionality to maximize the capture of a tracked participant or wearable item (e.g., thebracelet508 orshoestring tag516 described below) that is paired with the programmeddevice120.
As illustrated inFIG. 2A, the programmeddevice120 initially displays alogin interface48. In an embodiment, thelogin interface48 includes alogin element50. After the user activates thelogin element50, the programmeddevice120 displays theuser profile interface52 illustrated inFIG. 2B. As shown, theuser profile interface52 enables the user to create login credentials (e.g., username and password), enter personal information (e.g., cell phone number, email address and zip code), select a preferred language (e.g., English) and select a preferred temperature standard (e.g. English).
Once logged-in, the programmeddevice120 displays thehome interface54 as illustrated inFIG. 3A. Thehome interface54 displays a plurality ofcompilation videos60,61,62 and other compilation videos, below thecompilation video62, that are visible via swiping. As described further below, thecompilation videos60,61,62 haveratings63,65,67, respectively. The programmeddevice120 is operable to sort the videos, by default, according to the ratings such that the video with the highest rating is displayed at the top of thehome interface54. In an embodiment, the ratings represent likeness or flame per view, as described below.
In addition, thehome interface54 includes a plurality of icons or symbols at the bottom of thehome interface54. In the example shown, thehome interface54 displays ahome symbol72 that, upon selection, causes the programmeddevice120 to display thehome interface54. Thehome interface54 also displays aparticipant map symbol74, apeople follower symbol76 enabling the user to search for, select and follow other users (e.g., athletes or participants), avideo camera symbol78, and aconnection symbol80, each of which is described below.
It should be appreciated that thehome interface54 can be a mobile app interface, a website, or another online or network-accessible portal or medium, including, but not limited to, a social media, cloud-based platform. For example, thehome interface54 can be the front interface of the YouTube™ online video platform.
As illustrated inFIGS. 2B and 3A, the programmeddevice120 also displays amenu element81. In response to the user's selection or activation of themenu element81, the programmeddevice120 displays afeatures interface82 as illustrated inFIG. 3B. The features interface82 displays a plurality of functions of thesystem13. In the example shown, thefeatures interface82 displays: (a) ahome element84 selectable by the user, which serves the same function as thehome symbol72; (b) a user profile element86 selectable by the user, enabling the user to log-out or change user accounts; (c) a filming options or videorecording options element88; (d) aparticipant center element90; and (e) aconnector element92, which serves the same function as theconnection symbol80.
In the embodiment shown inFIG. 3A, thehome interface54 displays asearch interface312. Thesearch interface312 displays afilter switch95, anupdate filter element97, atext search field99, asearch activator101 and afollower search element103. The sliding of thefilter switch95 to the left (corresponding to “all”) effectively turns-off the search filter. The sliding of thefilter switch95 to the right (corresponding to “my filter”) effectively turns-on the search filter.
Also, the user can select theupdate filter element97. In response to the user's selection of theupdate filter element97, the programmeddevice120 displays theupdate filter interface105 as illustrated inFIG. 3C. Theupdate filter interface105 displays anevent selector107, agender selector109, aminimum age selector111, amaximum age selector113, alocation field115, aproximity field117 and asave filter element119. Referring toFIG. 4, the programmeddevice120 displays: (a) an event descriptor category, event reel orevent strip121 in response to the user's selection of theevent selector107; (b) a gender descriptor category, a gender reel orgender strip123 in response to the user's selection of thegender selector109; (c) a minimum age descriptor category, a minimum age reel or aminimum age strip125 in response to the user's selection of theminimum age selector111; and (d) a maximum age descriptor category, a maximum age reel or amaximum age strip127 in response to the user's selection of themaximum age selector113. In the example shown, theevent strip121 displays a strip of elements associated with different types of events, including abaseball element96,basketball element98,football element100,soccer element102,martial arts element104, track andfield element106, science technology engineering and math (STEM) element107 (associated with presentations at science fairs and other STEM venues), business presentation element109 (associated with business plan/investor pitch competitions), and ageneral element111 associated with any other type of non-categorized event, including, but not limited to, any sport or non-sport activity, such as debate club, acting, music, dancing and other activities.
In response to the user's selection of one of these event elements, thesystem13 changes the event element to correspond to the selected event element. In the example shown, the user selectedbasketball element102, the programmeddevice120 highlighted thebasketball element98, and the programmeddevice120 displayed thebasketball element98 at the top of theevent strip121. In response to the user's selection of one of the gender elements, thesystem13 changes the gender element to correspond to the selected gender element. In the example shown, the user selectedfemale element131, the programmeddevice120 highlighted thefemale element131, and the programmeddevice120 displayed thefemale element131 at the top of thegender strip123. In response to the user's selection of one of the minimum age elements, thesystem13 changes the minimum age element to correspond to the selected minimum age element. In the example shown, the user selected minimum age fifteen, the programmeddevice120 highlighted the numeral fifteen, and the programmeddevice120 displayed the numeral fifteen at the top of theminimum age strip125. In response to the user's selection of one of the maximum age elements, thesystem13 changes the maximum age element to correspond to the selected maximum age element. In the example shown, the user selected maximum age seventeen, the programmeddevice120 highlighted the numeral seventeen, and the programmeddevice120 displayed the numeral seventeen at the top of themaximum age strip127. Accordingly, in this example, the user set a custom filter for videos that involve basketball and female participants (i.e., female basketball players) having an age within the range of fifteen to seventeen years old. The update filter interface105 (FIG. 3C) then indicates the user's filter setting and provides the user with the opportunity to narrow the search further by: (a) entering a location (e.g., city, zip code, state or country) in the location descriptor category orlocation field115; and/or (b) entering a radial distance in the proximity descriptor category orproximity field117, such as twenty-five miles or kilometers from such location. In response to the user's selection of thesave filter element119, thesystem13 saves the filter setting indicated by theupdate filter interface105.
It should be appreciated that thesearch interface312 can include or be operatively coupled to a plurality of descriptor categories other than those illustrated inFIGS. 3A-4, including, but not limited to, country, city, state, language, race, ethnicity, school name, grade point average (“GPA”), ACT score, SAT score, coach's name, position, height, weight, shooting percentage, points per game, other performance statistics, and other types of participant characteristics.
Returning to the home interface54 (FIG. 3A), if the user swipes thefilter switch95 to the right, the programmeddevice120 displays thecompilation videos60,61,62 according to the filter setting indicated by theupdate filter interface105. If the user swipes thefilter switch95 to the left, the programmeddevice120 displays thecompilation videos60,61,62 without any filtering. If the user enters text in the text search field99 (e.g., an athlete's name) and then selects thesearch activator101, the programmeddevice120 processes a search request and displays thecompilation videos60,61,62 according to the text entered in thetext search field99. If the user selects thefollower search element103, the programmeddevice120 blocks or deactivates any filter settings and displays thecompilation videos60,61,62 of those users who are followed by the user in accordance with the settings input through thepeople follower symbol76.
As illustrated inFIGS. 5A-5B, in response to the user's selection of theparticipant map symbol74, thesystem13 displays themap interface108. Themap interface108 displays asearch field110 that enables the user to enter a zip code or name of city, state or other territory. Upon entering the data in the field110 (e.g.,zip code 60426 of Harvey, Ill.), thesystem13 displays ageographic map94 of users who are registered through thesystem13 as participants. In an embodiment, thegeographic map94 graphically represents participants according to the update filter interface105 (FIG. 3C). The map displays symbols or different sizes, shapes or colors to indicate the athletes of varying ratings. In the example shown, the relatively small squares indicate athletes with ratings below a designated level, and the three relatively large squares indicate athletes with ratings above the designated level. In response to the user's selection of one of the symbols, thesystem13 displays biographical information regarding the corresponding athlete. In the example shown, the user enteredzip code 60426 of Harvey, Ill. for a search for high school female basketball players, and themap interface108 displayed a map of Harvey, Ill. populated with the locations or school addresses of high school female basketball players indicated by squares. Next, the user selected the large central square, and themap interface108 displayed information regarding the corresponding basketball player—Tyra Wilson, 6′ point guard,age 16, Thornton High School, Harvey, Ill.
The search interface312 (FIG. 3A) and the map interface108 (FIGS. 5A-5B) overcome challenges and barriers encountered by participants, such as athletes aspiring to play sports in college. For example, it is common for talented high school athletes to be overlooked because they attend low profile high schools, reside in relatively small cities or towns, do not satisfy the ideal height and weight for a given sport, lack the personal connections, or lack the financial resources to pay recruiting consultants. These athletes, who play on high school and Amateur Athletic Union (“AAU”) teams, often find it difficult to gain adequate exposure to recruiters, colleges, teams and media.
Using conventional (prior art) video platforms like YouTube™, it can be difficult, burdensome and time consuming for recruiters and sports enthusiasts to identify athletes who match a desired profile, such as age, gender, sport type, performance statistic, height, weight, GPA or other descriptors of various descriptor categories. For example, a YouTube™ search for “top 17 year old high school girl basketball players in Cleveland, Ohio” may result in 83,900 results with the first five including: (a) The Best High School Basketball Player From Every State; (b) 7′7 freshman makes varsity debut; (c) 7-Foot-7 190 lbs Freshman; (d) 7′7″ basketball player in Ohio; and (e) Chargrin Falls' senior Hallie Thome named Cleveland.com's Girls Basketball Player of the Year. Four of the top five results do not even involve girl basketball players, and the fifth result involves a eighteen year old girl basketball player. The sought-after player may be buried in the 83,900 results, requiring searchers to spend hours to identify 17 year old girl basketball players in Cleveland, Ohio. Thesystem13 provides an improvement that overcomes or decreases the effects of this problem. In particular, the search interface312 (FIG. 3A) enables users to use thefilter95 to find compilation videos of participants that satisfy the specific descriptors selected by the users. In an embodiment described below, thesystem13 requires the video submitter to input descriptors, such as event type, gender, age and zip code, into the primary video categorizer interface287 (FIG. 40A).
Themap interface108 enables recruiters to conveniently investigate the athletes within a desired geography. For example, without themap interface108, recruiters might avoid traveling to a small town to view a single athlete. With the improvement and advantage provided by themap interface108, a recruiter can virtually visit small towns and view the videos and information regarding the athletes there. In addition, as described above, the search interface31 (FIG. 3) enables recruiters to filter and narrowly search for athletes and participants who satisfy specific criteria input by the recruiters. This functionality and the advantages of theconnector module36 described below, provide important improvements that overcome or lessen the disadvantages described above.
As illustrated inFIG. 6A, when a user selects the recording options element88 (FIG. 3B), the programmeddevice120 displays therecording options interface110. The recording options interface110 displays astandard mode element112, custom mode element114,standard cutback116, custom cutback field118,standard cutforward120, custom cutforward field122, and recording featureselement124.
If the user selects thestandard mode element112, the programmeddevice120 automatically activates thestandard cutback116 andstandard cutforward120. Thestandard cutback116 andstandard cutforward120 are the default values. In the example shown, the value of thestandard cutback116 is set at five seconds, and the value of the standard cutforward122 is set at two seconds. It should be appreciated that these values can be adjusted by the implementor of thesystem13.
If the user selects the manual mode element114, the programmeddevice120 deactivates thedefault cutback116 anddefault cutforward120, and the programmeddevice120 enables the user to enter the desired data (e.g., time values in seconds) in the custom cutback field118 and custom cutforward field122. As described further below, the time values established in the recording options interface110 affect the video clipping process.
In response to the user's selection of the recording featureselement124, the programmeddevice120 displays the recording featuresinterface126 as illustrated inFIG. 6B. In an embodiment, the recording featuresinterface126 displays: (a) abasic mode element128; (b) anadvanced mode element130; (c) ahighlights element132 associated with success or positive activity of a participant's performance; (d) alowlights elements134 associated with failure, weakness or negative activity that indicates areas for training or improvement in a participant's skills; and (e) astats element136 associated with a set of statistics symbols216 (FIG. 26) described below.
In response to the user's selection of thebasic mode element128, thesystem13 activates abasic recording mode140 as illustrated inFIG. 7. According to the basic method of use indicated inFIG. 7:
- (a) To activate the recording function of the programmeddevice120, the user presses or taps thevideo camera symbol78 as illustrated inFIGS. 3A and 6A-6B. In response, the programmeddevice120 displays a recording interface142 as illustrated inFIG. 8.
- (b) To start recording, the user presses and holds the start/stop element144 (FIG. 7) which, in the example shown, is a wheel symbol. After the user continuously presses the start/stop element144 for a designated period (e.g., one second), the programmeddevice120 animates the start/stop element144 and starts the recording of the event. In the example shown, the programmeddevice120 causes the wheel symbol to spin or rotate. The continuous motion of the wheel symbol indicates that recording is in progress. It should be appreciated that, in other embodiments, the start/stop element144 can include other animated symbols, such as a spinning basketball, spinning football, spinning baseball, spinning soccer ball, another spinning or moving sports object associated with a particular sport, or a dot or ball that travels clockwise around the perimeter (the path of flash150).
- (c) To capture video footage146 (FIG. 8) of the recorded event, the user presses and holds one or more fingers (or another part of the user's body) on the touchscreen148 (FIG. 9) of the programmeddevice120 until thesystem13 displays a relatively bright flash150 (FIG. 10A) located at the perimeter of the recording interface142. In this embodiment, the programmeddevice120 has a designated confirmation period, such as two seconds. The programmeddevice120 checks to determine whether the user has made a continuous, intentional input onto the touchscreen148 for the confirmation period. Once the programmeddevice120 confirms that the user has satisfied this condition, the programmeddevice120 proceeds to generate theflash150 and capture thevideo footage146. It should be appreciated that, in other embodiments, the programmeddevice120 is configured to receive other types of actions or inputs to generate the desiredvideo footage146, including, but not limited to, voice, audible, retinal, biometric and gesture inputs, user actions, movements of the programmeddevice120 relative to other objects, and electronic signals from ancillary devices, sensors or accessories. The flash150 (FIG. 10A) indicates to the user that the programmeddevice120 has successfully received the user's input to generate the desiredvideo footage146. In an embodiment, theflash150 is bright white, silver, yellow, orange or red. In another embodiment, theflash150 is a graphical animation of a rectangular path or line of fire showing a line of red and orange flames in motion. In yet another embodiment, the programmeddevice120 displays a sequence offlashes150 in which flash150 quickly changes between illuminated and non-illuminated appearances. After the flashing or flash period ends, the programmeddevice120 deactivates theflash150, returning to the recording interface142 shown inFIG. 10B.
- (d) To pause or stop the recording, the user presses and holds the start/stop element144. After the user continuously presses the start/stop element144 for a designated period (e.g., one second), the programmeddevice120 stops the animation of the start/stop element144 and stops the recording of the event. In the example shown, the programmeddevice120 stops the spinning and rotation of the wheel symbol. The stationary display of the wheel symbol indicates that recording has stopped or paused.
- (e) To wrap-up, end or terminate the recording session, the user presses or selects therecording exit element145. In addition, the user can use his/herhand152 to cover therear camera lens154 of the programmeddevice120 as illustrated inFIGS. 11-12. The programmeddevice120 checks to determine whether the user has made a continuous, intentional covering of thelens154 for a confirmation period, such as one second. Once the programmeddevice120 confirms that the user has satisfied this condition, the programmeddevice120 recognizes an exit input. In an embodiment, in response to an exit input through theexit element145 or therear camera lens154, the programmeddevice120 automatically displays a publishdecision interface156 as illustrated inFIG. 13A. The publishdecision interface156 displays a continue recordingelement158 and a publish nowelement160. Depending upon the embodiment, the publishdecision interface156 can cover or replace the entire recording interface142, or the publishdecision interface156 can be a pop-up window that overlays only part of the recording interface142. If the user selects the continue recordingelement158, the programmeddevice120 displays the recording interface142. If the user selects the publish nowelement160, the programmeddevice120 automatically publishes a highlight video having a compilation of select video clips of thevideo footage146, or the programmeddevice120 enables the user to add information before publishing such video, as described further below. The publishdecision interface156 provides a secondary safeguard against an unintentional stoppage of recording. The confirmation period for the lens covering can serve as a primary safeguard.
In response to the user's selection of the advanced mode element130 (FIG. 6B), the programmeddevice120 activates anadvanced recording mode162 as illustrated inFIGS. 14-15. According to the advanced method of use described inFIGS. 14-15:
- (a) To activate the recording function of the programmeddevice120, the user presses or taps thevideo camera symbol78 as illustrated inFIGS. 3A and 6A-6B. In response, the programmeddevice120 displays a recording interface142 as illustrated inFIG. 16.
- (b) To start recording, the user presses and holds the start/stop element144 (FIG. 16) which, in the example shown, is a wheel symbol. After the user continuously presses the start/stop element144 for a designated period (e.g., one second), the programmeddevice120 animates the start/stop element144 and starts the recording of the event. In the example shown, the programmeddevice120 causes the wheel symbol to spin or rotate. The continuous motion of the wheel symbol indicates that recording is in progress. It should be appreciated that, in other embodiments, the start/stop element144 can include other animated symbols, such as a spinning basketball, spinning football, spinning baseball, spinning soccer ball or another spinning or moving sports object associated with a particular sport.
- (c) To generate or capture a video clip while, at the same time, recording the statistic associated with the video clip, the user provides one of the clip-stat commands164 (FIG. 14), multi-functional commands. As shown inFIG. 15, the programmeddevice120 stores a plurality ofcorrelations166 related to the clip-stat commands164.
- (d) As illustrated inFIG. 16, if the user presses or taps one finger at anysingle spot168 on the touchscreen148, this single-finger input has a one input characteristic associated with a scoring of one point (e.g., a basketball free throw or soccer goal). This causes the programmeddevice120 to simultaneously save or record one point and generate or capture the associated video clip, as described below. In an embodiment illustrated inFIGS. 14 and 16, if the user presses or taps one finger at anysingle spot168 on the touchscreen148, the programmeddevice120 simultaneously: (i) saves or records one point; (ii) generates or captures the associated video clip, as described below; and (iii) displays a statistics capture confirmation, such as a “1” appearing momentarily on the touchscreen148 and then disappearing as indicated inFIG. 14.
- (e) As illustrated inFIG. 17, if the user simultaneously presses or taps two fingers at any twospots170,172 on the touchscreen148, this two-finger input has a two input characteristic associated with a scoring of two points (e.g., a basketball field goal). This causes the programmeddevice120 to simultaneously save or record two points and generate or capture the associated video clip, as described below. In an embodiment illustrated inFIGS. 14 and 17, if the user simultaneously presses or taps two fingers on any twospots172,174 on the touchscreen148, the programmeddevice120 simultaneously: (i) saves or records two points; (ii) generates or captures the associated video clip, as described below; and (iii) displays a statistics capture confirmation, such as a “2” appearing momentarily on the touchscreen148 and then disappearing as indicated inFIG. 17.
- (f) As illustrated inFIG. 18, if the user simultaneously presses or taps three fingers at any threespots174,176,178 on the touchscreen148, this three-finger input has a three input characteristic associated with a scoring of three points (e.g., a basketball field goal behind the three point arc). This causes the programmeddevice120 to simultaneously save or record three points and generate or capture the associated video clip, as described below. In an embodiment illustrated inFIGS. 14 and 18, if the user simultaneously presses or taps three fingers on any threespots174,176,178 on the touchscreen148, the programmeddevice120 simultaneously: (i) saves or records three points; (ii) generates or captures the associated video clip, as described below; and (iii) displays a statistics capture confirmation, such as a “3” appearing momentarily on the touchscreen148 and then disappearing as indicated inFIG. 14.
- (g) As illustrated inFIG. 19, if the user laterally drags or swipes one or more fingers from left to right or right to left on the touchscreen148 along a lateral or substantiallylateral path180, the lateral swiping input has a lateral or horizontal input characteristic associated with a lateral or horizontal path of a passed ball (e.g., the passing of a basketball from one player to another player who scores). In an embodiment, this lateral or horizontal input characteristic is associated with the passing or movement of a ball or sports object substantially laterally or horizontally across a court or sports area. In basketball, the user could provide this input when a player passes a ball that results in an assist. This input causes the programmeddevice120 to simultaneously save or record one assist and generate or capture the associated video clip, as described below. In an embodiment illustrated inFIGS. 14 and 19, if the user drags one or more fingers along substantiallylateral path180, the programmeddevice120 simultaneously: (i) saves or records one assist; (ii) generates or captures the associated video clip, as described below; and (iii) displays a statistics capture confirmation, such as a “ASSIST” appearing momentarily on the touchscreen148 and then disappearing as indicated inFIG. 14.
- (h) As illustrated inFIG. 20, if the user vertically drags or swipes one or more fingers upward on the touchscreen148 along an upward or substantiallyupward path182, the upward swiping input has a rise, jumping, vertical or upward input characteristic associated with the substantiallyupward path182 of the rising motion of a player jumping upward (e.g., the upward jumping of a basketball player to rebound a ball). In an embodiment, this upward input characteristic is associated with the rebounding of a ball or sports object. In basketball, the user could provide this input when a player successfully rebounds a ball. This input causes the programmeddevice120 to simultaneously save or record one rebound and generate or capture the associated video clip, as described below. In an embodiment illustrated inFIGS. 14 and 20, if the user drags one or more fingers along the substantiallyupward path182, the programmeddevice120 simultaneously: (i) saves or records one rebound; (ii) generates or captures the associated video clip, as described below; and (iii) displays a statistics capture confirmation, such as a “REBOUND” or a symbol thereof appearing momentarily on the touchscreen148 and then disappearing as indicated inFIG. 14.
- (i) As illustrated inFIG. 21, if the user simultaneously presses or taps all four fingers (and optionally, the thumb) at any fourspots184,186,188,190 or more on the touchscreen148, this four-finger input has a hand input characteristic associated with an entire hand that is typically involved in stealing a ball from an opponent (e.g., a steal in basketball). This input causes the programmeddevice120 to simultaneously save or record one steal and generate or capture the associated video clip, as described below. In an embodiment illustrated inFIGS. 14 and 21, if the user simultaneously presses or taps four fingers on any fourspots184,186,188 and190 on the touchscreen148, the programmeddevice120 simultaneously: (i) saves or records one steal; (ii) generates or captures the associated video clip, as described below; and (iii) displays a statistics capture confirmation, such as “STEAL” or a symbol thereof appearing momentarily on the touchscreen148 and then disappearing as indicated inFIG. 4.
- (j) As illustrated inFIG. 22, if the user simultaneously presses or taps the palm orbase192 of a first at anyspot194 on the touchscreen148, this large surface or fist-shaped input has a powerful or protective input characteristic associated with a fight or action to block or reject an opponent (e.g., a block in basketball). This input causes the programmeddevice120 to simultaneously save or record one block and generate or capture the associated video clip, as described below. In an embodiment illustrated inFIGS. 14 and 22, if the user simultaneously presses or taps thebase192 of the hand on anyspot194 on the touchscreen148, the programmeddevice120 simultaneously: (i) saves or records one block; (ii) generates or captures the associated video clip, as described below; and (iii) displays a statistics capture confirmation, such as “BLOCK” or a symbol thereof appearing momentarily on the touchscreen148 and then disappearing as indicated inFIG. 14.
- (k) As illustrated inFIG. 23, if the user vertically drags or swipes one or more fingers to draw an X by swiping along intersectingpaths196,198, the X-shaped input has a cancel, error or negative input characteristic associated with a mistake, error or underperformance of a player (e.g., a turnover by a basketball player). In an embodiment, this negative input characteristic is associated with a turnover caused by a basketball, football or other athlete. In basketball, the user could provide this input when a player loses the ball or otherwise performs a turnover. This input causes the programmeddevice120 to simultaneously save or record one turnover and generate or capture the associated video clip, as described below. In an embodiment illustrated inFIGS. 14 and 23, if the user drags one or more fingers along the intersectingpaths196,198, the programmeddevice120 simultaneously: (i) saves or records one turnover; (ii) generates or captures the associated video clip, as described below; and (iii) displays a statistics capture confirmation, such as a “TURNOVER” or a symbol thereof appearing momentarily on the touchscreen148 and then disappearing as indicated inFIG. 14.
There are several challenges and difficulties that event attendees encounter when video recording events (e.g., games) while, at the same time, trying to document important statistics regarding the events. First, the attendee experiences a series of emotional rises and falls throughout the event. Often, the pivotal moments in the event can cause the attendee to momentarily lose attention of the video recording or statistics. These emotions increase the difficulty to reliably video record all of the important footage of a designated player while reliably recording all of the important statistics of such player.
The programmeddevice120 overcomes or substantially decreases this difficulty by providing several technical advantages. As described further below, thevideo generator28 of the programmeddevice120 has a clipping logic that enables the attendee to capture important footage after the pivotal moments have occurred. This avoid the burden of trying to remember to cut or clip pivotal moments while the moments are occurring. Also, thecorrelations166 of theadvanced recording mode162, described above, enable the attendee to seamlessly capture a video clip and the associated statistic at the same time based on a single input. In addition, the characteristic of the input resembles or relates to the statistic. For example, a tap of one finger relates to a statistic of one point. This provides a cognitive learning and memory advantage by making it easier to remember which type of input to provide for a given statistic. This enhanced human machine interface simplifies the overall process of capturing important video clips and recording important statistics related to the video clips.
In another embodiment illustrated inFIGS. 24A-25B, the programmeddevice120 generates arecording interface202 in response to the user's activation of the video camera symbol78 (FIG. 3A). Therecording interface202 includes a start/stop element204, a wrap-up orexit element206, ahighlight clip element208 and alowlight clip element210. The start/stop element204 includes an on indicator, such as an illuminated or colored graphic as well as a timer. In the example shown inFIG. 24A, the start/stop element204 is a basketball symbol, and once the user presses or taps the start/stop element204, the perimeter of the basketball symbol has an illuminated orange circle or arc, and the timer continuously increments from 0:00 to 0:01 to 0:02 to 0:03 and eventually to 1:19 and onward. To generate or capture a video clip of important, positive footage (e.g., a score, steal, assist, rebound or other highlight212) the user can press or tap thehighlight clip element208. In the example shown, thehighlight clip element208 is a fire symbol. To generate or capture a video clip of important, negative footage (e.g., a turnover, missed shot, error, mistake, blunder, underperformance, inappropriate behavior of a coach, or other lowlight) the user can press or tap thelowlight clip element210. In the example shown, thelowlight clip element210 is an ice or icicle symbol. When the user is ready, such as at the end of the game, the user can press or tap the wrap-up orexit element206. In response, the programmeddevice120 displays the publish decision interface156 (FIG. 13A) which, in turn, displays the continue recordingelement158 and publish nowelement160, as described above.
In another embodiment illustrated inFIG. 26, the programmeddevice120 generates arecording interface214 in response to the user's activation of the video camera symbol78 (FIG. 3A). In this embodiment, therecording interface214 displays a set ofstatistics symbols216. In the basketball example shown, thestatistics symbols216 include a threepoint symbol218, a twopoint symbol220, a free throw (one point)symbol222, anassist symbol224, ablock symbol226, arebound symbol228, asteal symbol230, and aturnover symbol232.
In an embodiment, therecording interface214 enables the user to generate video clips while recording statistics through use of thestatistics symbols216. Depending upon the embodiment, the recording interface214: (a) displays the solid images of thestatistics symbols216 on top of the recorded imagery; or (b) displays the translucent or partially transparent images of thestatistics symbols216 on top of the recorded imagery.
In an embodiment, therecording interface214 includes and displays a statistics icon (not shown), such as an image of a clipboard or statistics book. During the recording session, therecording interface214 displays such statistics icon, and the default is to hide (or otherwise not display) thestatistics symbols216. When the user presses the statistics icon, therecording interface214 displays or pops-up thestatistics symbols216. This enables the user to select theappropriate statistics symbols216 to record the applicable statistic.
In various embodiments described above, the type of inputs from the user to the programmeddevice120 involves a touching or tapping of the touchscreen148. It should be appreciated that, in other embodiments, the user can provide alternate types of inputs. In such embodiments, it is not necessary for the programmeddevice120 to have a touchscreen148.
In an embodiment, thesystem13 enables the programmeddevice120 to receive audio or sound inputs for voice commands. In a setup process, the programmeddevice120 enables the user to train the programmeddevice120 to recognize sound signatures or unique voice sounds produced by the user. For example, the user can output different oral statements into the microphone of the programmeddevice120. The oral statements corresponds to different types of statistics, such as “ONE,” “TWO,” “THREE,” “ASSIST,” “REBOUND,” “STEAL,” “BLOCK,” and “TURNOVER.”
In this embodiment, the programmeddevice120 includes a comparator that compares the user's unique voice to the environmental sounds, such as the roars of the crowd and voice commands of other attendees in the audience who are using programmeddevices120 on their electronic devices. The comparator identifies the user's voice so that the programmeddevice120 does not register non-user sounds as voice commands by the user. In an embodiment, the programmeddevice12 includes a sound confusion inhibitor that enables the user to record a unique voice activation sound, such as the first name, last name, initial or jersey number of the particular player for which the user is recording statistics. For example, the voice activation sound could be “JOHN,” JUSTICE” or “J.” In such example, the oral statements corresponding to the different types of statistics could be as follows: “J ONE,” “J TWO,” “J THREE,” “J ASSIST,” “J REBOUND,” “J STEAL,” “J BLOCK,” and “J TURNOVER.” If the user does not speak “J” before speaking the applicable statistic, thesystem13 will not record such statistic.
In an embodiment, the programmeddevice120 displays a pop-up or confirmation of the recorded statistic to confirm the statistic that the user input through his/her voice. For example, thesystem13 can cause the programmeddevice120 to display “ONE POINT” by itself or “ONE POINT” adjacent to a garbage symbol, in which case the user can press the garbage symbol if such statistic is wrong. If the user taps the garbage symbol, the programmeddevice120 discards or otherwise does not record such erroneous statistic.
In another embodiment, the programmeddevice120 enables the user to provide inputs through physical interaction with the programmeddevice120, such as by applying forces to the programmeddevice120, accelerating or moving the programmeddevice120 or changing the orientation or position of the programmed device120 (e.g., rotating or twisting the programmed device120). In such embodiment, the programmeddevice120 includes one or more sensors (including, but not limited to, accelerometers) configured to sense or detect forces, light changes, movement or positional change of the programmeddevice120. For example, to start or stop a recording session, thesystem13 can enable the user to quickly turn the programmeddevice120 face up (to start) or face down (to stop). In another example, thesystem13 can enable the user to record inputs for different statistics by: (a) sharply tapping the back case of the programmeddevice120 one time to record one point; (b) sharply tapping the back case of the programmeddevice120 two times to record two points; and (c) sharply tapping the back case of the programmeddevice120 three times to record three points.
As described above, the recording options110 (FIG. 6A) enable the user to select the default orstandard cutback116 andcutforward120 or to input a custom cutback118 and custom cutforward122. The user can, for example, input ten seconds for the custom cutback118. If the user selects the standard cutback116 (e.g., five seconds), thevideo generator28 reaches backward five seconds to initiate the cut for the applicable video clip, as described below.
In an embodiment, when the user provide an input to generate a video clip, the programmeddevice120 displays a cutback pop-up234 as illustrated inFIG. 27. This enables the user to switch to the custom cutback118 on a case-by-case basis. For example, a player may have been involved in action that lasted for a relatively long period, such as a75 yard run by a football player or a basketball player's steal, then turnover, then recovery of the ball, then drive and dunk. If the user encounters such lengthy action, the user may desire to tap the cutback pop-up234. In response, the programmeddevice120 will cut the beginning of the clip, ten seconds before the time of the user's clip input.
Referring toFIGS. 28-30B, in an embodiment, theelectronic device120 generates a video through a clipping process. During the recording session, thevideo generator28 of the programmeddevice120 is operable to generate adata list236. Also, during the recording session, the programmeddevice120 generates a video track238 (FIGS. 30A-30B) over a period of time.
In the examples described, the time increments are seconds. It should be appreciated, however, that the time increments can be milliseconds or any other suitable increment. Also, the programmeddevice120 is operable to generate and store thevideo track238 through a rate capture rate within the range of thirty to one thousand frames per second (FPS) or through a rate capture rate of any other suitable FPS.
In the example shown, once the recording session starts, the programmeddevice120 generates and stores a continuous stream, track or series of timestamps in chronological order based on a suitable time increment. In the example shown, the increment is seconds, and the programmeddevice120 generated timestamps one through twenty-three. In this example, the user provided a first clip input at the point of twelve seconds, as indicated by the first arrow A1 shown inFIG. 30B. In response, the programmeddevice120 flagged, marked or bookmarked the twelve second point by storing a suitable data marker A1 (FIG. 28), which corresponds to the first clip input. At the same time or thereafter, the programmeddevice120 flagged, marked or bookmarked the seven second point by storing a suitable data marker A2 (FIG. 28), which corresponds to the first rearward point. Later, the user provided a second clip input at the point of twenty seconds, as indicated by the second arrow A3 shown inFIG. 30B. In response, the programmeddevice120 flagged, marked or bookmarked the twenty second point by storing a suitable data marker A3 (FIG. 29), which corresponds to the second clip input. At the same time or thereafter, the programmeddevice120 flagged, marked or bookmarked the fifteen second point by storing a suitable data marker A4 (FIG. 29), which corresponds to the second rearward point.
As illustrated inFIG. 30B, thevideo track238 includes a video clip X1 between the data markers A2 and A1, and thevideo track238 includes a video clip X2 between the data markers A4 and A3. In this example, during the recording session the programmeddevice120 automatically cut-out and deleted theexcess tracks240,242 from thevideo track238, and the programmeddevice120 automatically deleted theexcess track240 before recording theexcess track242. As described above, this helps preserve data storage capacity on the programmeddevice120. In an embodiment, the programmeddevice120 automatically deletes theexcess track240 immediately in response to the first clip input at A1, and the programmeddevice120 automatically deletes theexcess track242 immediately in response to the second clip input at A3. In other embodiments, as described below, the programmeddevice120 deletes the excess tracks after the recording session ends, not during the recording session.
In another embodiment, the clipping process involves look-rearward and look-forward steps. In the example shown inFIGS. 30-33, once the recording session starts, thevideo generator28 of programmeddevice120 is operable to generate adata list244. Thevideo generator28 stores a continuous stream, track or series of timestamps in chronological order based on a suitable time increment. In the example shown, the increment is seconds, and the programmeddevice120 generated timestamps one through twenty-three. In this example, the user provided a first clip input at the point of ten seconds, as indicated by the first arrow B1 shown inFIG. 33. In response, the programmeddevice120 flagged, marked or bookmarked the ten second point by storing a suitable data marker B1 (FIG. 31), which corresponds to the first clip input. At the same time or thereafter, the programmeddevice120 flagged, marked or bookmarked the five second point by storing a suitable data marker B2 (FIG. 31), which corresponds to the first rearward point. Simultaneously or a moment thereafter, the programmeddevice120 flagged, marked or bookmarked the twelve second point by storing a suitable data marker B3 (FIG. 31), which corresponds to the first forward point.
Later, the user provided a second clip input at the point of twenty seconds, as indicated by the second arrow B4 shown inFIG. 33. In response, the programmeddevice120 flagged, marked or bookmarked the twenty second point by storing a suitable data marker B4 (FIG. 32), which corresponds to the second clip input. At the same time or thereafter, the programmeddevice120 flagged, marked or bookmarked the fifteen second point by storing a suitable data marker B5 (FIG. 32), which corresponds to the second rearward point. Simultaneously or a moment thereafter, the programmeddevice120 flagged, marked or bookmarked the twenty-two second point by storing a suitable data marker B6 (FIG. 33), which corresponds to the second forward point.
As illustrated inFIG. 33, thevideo track238 includes a video clip X2 extending continuously between the data markers B2 and B3, and thevideo track238 includes a video clip X3 extending continuously between the data markers B6 and B5. In this example, during the recording session the programmeddevice120 automatically cut-out and deleted theexcess tracks246,248 from thevideo track238, and the programmeddevice120 automatically deleted theexcess track246 before recording theexcess track248. As described above, this helps preserve data storage capacity on the programmeddevice120. In other embodiments, as described below, the programmeddevice120 deletes the excess tracks after the recording session ends, not during the recording session.
In another embodiment, the clipping process involves interference management in addition to the look-rearward and look-forward steps described above. In the example shown inFIGS. 34-36, once the recording session starts, thevideo generator28 of programmeddevice120 is operable to generate adata list250. Thevideo generator28 stores a continuous stream, track or series of timestamps in chronological order based on a suitable time increment. In the example shown, the increment is seconds, and the programmeddevice120 generated timestamps one through twenty-three. In this example, the user provided a first clip input at the point of ten seconds, as indicated by the first arrow C1 shown inFIG. 36. In response, the programmeddevice120 flagged, marked or bookmarked the ten second point by storing a suitable data marker C1 (FIG. 34), which corresponds to the first clip input. At the same time or thereafter, the programmeddevice120 flagged, marked or bookmarked the five second point by storing a suitable data marker C2 (FIG. 36), which corresponds to the first rearward point. At the same time or thereafter, the programmeddevice120 flagged, marked or bookmarked the twelve second point by storing a suitable data marker C3 (FIG. 36), which corresponds to the first forward point.
Later, the user provided a second clip input at the point of fourteen seconds, as indicated by the second arrow C4 shown inFIG. 36. Notably, the second clip input occurs soon after the first clip input, only four seconds later. This could occur, for example, if the user provides a sequence of two or more clip inputs in rapid successions to capture separate, important moments, such as a football player's sacking of a quarterback, obtaining the football and then scoring a touchdown. Since the clip inputs occur close in time, the programmeddevice120 ensures that subsequent clip inputs do not interfere with previously captured video clips and do not cause the deletion of desired video clips.
Accordingly, in response to the second clip input at C4, the programmeddevice120 checks to determine whether any forward point timestamp has been marked that occurs in time less than five seconds before the second clip input C4. In this case, five seconds before C4 is the nine second point, and the first forward point C3 occurs at the twelve second point. Consequently, the programmeddevice120 uses the marker C3 as the data marker for the second rearward point. Therefore, the data marker C3 is associated with both a forward point and a rearward point. At the same time or thereafter, the programmeddevice120 flagged, marked or bookmarked the sixteen second point by storing a suitable data marker C5 (FIG. 36), which corresponds to the second forward point.
As illustrated inFIG. 36, thevideo track238 includes a video clip X4 extending continuously between the data markers C2 and C3, and thevideo track238 includes a video clip X5 extending continuously between the data markers C3 and C5. In this example, during the recording session the programmeddevice120 automatically cut-out and deleted theexcess track252 from thevideo track238, and the programmeddevice120 automatically deleted theexcess track252 after determining that the rearward point C2 is not the forward point of any previous video clip. As described above, in this example, the second clip input C4 did not cause the programmeddevice120 to delete any portion of video clip X4 because the programmeddevice120 determined that the rearward point C3 of the video clip X5 is the forward point C3 of video clip X4. An advantage of this interference management function is to safeguard against the undesirable deletion of video clips. In other embodiments, as described below, the programmeddevice120 deletes the excess tracks after the recording session ends, not during the recording session.
Referring toFIGS. 37-38, in an embodiment, the programmeddevice120 generates a video based on a bookmarking process. First, as indicated bystep254, the programmeddevice120 receives an input that starts the recording session, such as the user's tapping of the start/stop element144 (FIG. 8) or start/stop element204 (FIG. 24A). In this example, the user taps the start/stop element at the zero time point. As indicated bystep256, the programmeddevice120 then records the event (e.g., a basketball game or debate competition), and the programmeddevice120 continuously stores or saves the footage orvideo track238 as the event is being recorded. The programmeddevice120 can save thevideo track238 within a memory device component of the programmeddevice120, within a data storage disk operatively coupled to the programmeddevice120, or within a data storage device that is remote from the programmeddevice120, such as a webserver or data storage device12 (FIG. 1).
During the recording session, the programmeddevice120 determines whether the user has provided a stop input as indicated by thedecision diamond258. If the answer is yes, the programmeddevice120 pauses or stops the recording session, as indicated by thestep260, and then awaits for another start input as indicated by thestep254. If the answer is no, the programmeddevice120 continues the recording session.
During the recording session, the programmeddevice120 is operable to receive a plurality of different statistic inputs from the user as indicated bystep262. The programmeddevice120 stores the statistics (e.g., statistical data) associated with the statistic inputs. The programmeddevice120 can save the statistics within a memory device component of the programmeddevice120, within a data storage disk operatively coupled to the programmeddevice120, or within a data storage device that is remote from the programmeddevice120, such as a webserver or data storage device12 (FIG. 1).
Next, the programmeddevice120 receives a clip input at an input time point as indicated bystep264. Next, as indicated bystep266, the programmeddevice120 performs the following steps: (a) flags or bookmarks the input time point; (b) flags or bookmarks a rearward time point at R seconds (e.g., five seconds) before the input time point; and (c) flags or bookmarks a forward time point at F seconds (e.g., two seconds) after the input time point.
The automatic marking rearward in time and the automatic marking forward in time solve a pervasive problem experienced by typical users of prior art (conventional) recording devices. Users often miss important footage because they start or stop the video recording at the wrong times. For example, to save data storage capacity, users manually decide when to start and stop recording. When distracted, they often press the start button too late, so that the first part of the important footage is lost. Also, they often press the stop button too early, cutting off important footage. The programmeddevice120 solves this problem by enabling the user to continuously record, taking advantage of the auto-deletion function described below. While recording, the programmeddevice120 automatically captures the valuable moments by causing the clip marking to occur rearward and forward of the user's input time point.
Afterstep266, the programmeddevice120 determines whether the rearward time point precedes the forward time point of the previous video clip, if any, as indicated bydecision diamond268. This step is important to avoid the undesired deletion of previously saved video clips, as described above. If the answer is no, the programmeddevice120 proceeds to step270. If the answer is yes, the programmeddevice120 proceeds to step272.
The answer may be no because there were no previously saved video clips. Also, the answer may be no because the forward time point of the most recently saved video clip is before the rearward time point. In any case, if the answer is no, the programmeddevice120 automatically deletes the entire portion of thevideo track238 that occurs between the rearward time point and the forward time point of the most recent, preceding video clip as indicated bystep270. If there are no previously saved video clips, the programmeddevice120 automatically deletes the entire portion of thevideo track238 that occurs before the rearward time point.
The programmeddevice120 achieves several technical advantages by performing this auto-deletion function. Many events involve one or more relatively short, valuable actions or moments nested among dull, uninteresting or unimportant moments. For example, this is often the case for sports games, school debates, personal interviews and other events that are relatively long in duration. The prior art (conventional) process of editing a video after the recording is finished, can be time consuming, painstaking and burdensome. For example, producing a highlight video using the prior art process can take hours to edit the video tracks of an athlete's performance in a single game. Consequently, many videos with valuable moments are rarely viewed. People do not have the time or patience to watch long videos only to see a few valuable moments in the video. Nonetheless, for the sake of saving the valuable moments, users commonly save the full length of the videos on their prior art (conventional) mobile devices or on prior art (conventional) web servers.
This causes their prior art (conventional) mobile devices to reach maximum storage capacity, often in the midst of an event. Also, when users upload the full length videos to prior art (conventional) webservers, the webserver data centers consume substantial amounts of energy. For example, it has been reported that the data centers of Facebook®, YouTube® and others consume the equivalent of the energy output of numerous coal-fired power plants. Much of this energy goes to powering the webservers and keeping them cool. This energy is known to cause greenhouse gas emissions, resulting in a rising level of global pollution.
As described above, the auto-deletion function of thesystem13 helps free-up data storage capacity in electronic devices120 (e.g., smartphones) and in data storage devices12 (e.g., webservers). In an embodiment, while the user records an event, the programmeddevice120 purges or deletes the portions of the video track that contain dull, uninteresting or unvaluable footage. In such embodiment, the programmeddevice120 performs this deletion dynamically during and throughout the recording session. By automatically deleting the excess tracks during the recording session, the programmeddevice120 is less likely to reach maximum data storage capacity.
After thedeletion step270, the programmeddevice120 proceeds to step272. Atstep272, the programmeddevice120 retains or otherwise saves a video clip that is the portion of thevideo track238 between the rearward time point and the forward time point. Accordingly, the programmeddevice120 captures the applicable video clip of interest to the user. In an embodiment, the programmeddevice120 retains such video clip within thevideo track238 that is saved by the programmeddevice120. In another embodiment, the programmeddevice120 generates and saves a copy of such video clip and then deletes the original video clip from thevideo track238.
As the recording session continues, the programmeddevice120 receives another clip input at another input time point as indicated bystep274. Eventually, the user will be ready to end the recording session, such as at the end of the event. To do so, as indicated bystep276, the user provides a publish input or finish input by providing an input associated with the wrap-up, finalization or publication of a compilation video. Depending upon the embodiment, the user can provide this finish input by pressing the exit element145 (FIG. 8), covering the rear camera lens154 (FIG. 11), providing a sound input or providing another type of input.
In response to the finish input, the programmeddevice120 performs the following steps as indicated by step278: (a) combines and consolidates all of the saved video clips X1, X2, X3 (FIG. 38) in a chronological sequence with the first generated video clip occurring first, and the last generated video clip occurring last, resulting in a compilation video280 (FIG. 38); and (b) transfers the recorded stats to thepublication module31. Based on the auto-deletion function described above, the programmeddevice120 deleted the videotrack portions EXCESS1,EXCESS2, andEXCESS3 from thevideo track238. In an embodiment, thecompilation video280, such as a highlight video or so-called mixtape, has no blanks, null periods or blackout screens between the video clips X1, X2, X3. Thecompilation videos60,61,62 shown inFIG. 3A are videos, such ascompilation video280, produced by the programmeddevice120. As described below, the programmeddevice120 enables the user to add the recorded stats to a front video image of thecompilation video280.
It should be appreciated that, depending upon the embodiment, the programmeddevice120 can perform the auto-deletion function during or after the recording session. For example, in an embodiment, the programmeddevice120 deletes thetrack portions EXCESS1,EXCESS2, andEXCESS3 after the recording session ends in response to the finish input provided by the user. Such embodiment addresses the possibility that deleting the excess tracks during the recording session can overload or impair the processor ofprogrammed device120 depending upon the power of the processor. For example, by bookmarking during the recording without deleting, the processor of the programmeddevice120 will have more power availability to generate thevideo track238. By automatically deleting the excess tracks after the recording session, the programmeddevice120 is less likely to reach maximum data storage capacity during subsequent recording sessions.
As illustrated inFIG. 39, in response to the finish input, the programmeddevice120 generates processinginterfaces282,284,286. This indicates that the programmeddevice120 is in the process of generating thecompilation video280. Depending upon the embodiment, this process could take a fraction of second to several seconds. Next, referring toFIG. 40A, the programmeddevice120 generates the primaryvideo categorizer interface287 in accordance with the publication module31 (FIG. 1). The primaryvideo categorizer interface287 enables the user to enter a plurality of participant descriptors corresponding to a plurality of different descriptor categories, such as the event type, gender, age and zip code of or associated with the participant in the event. In response to thenext element289, the programmeddevice120 generates the secondaryvideo categorizer interface288 in accordance with thepublication module31 as illustrated byFIG. 40B. The secondaryvideo categorizer interface288 indicates a plurality of selectable video categories, such as Athlete Highlights, Athlete Development, Athlete Lowlights, AAU Team, Camp, College Recruiter, Physical Therapist, Sports Agent, Trainer and Tutor. In the example shown, the user selected the Athlete Highlights category.
In an embodiment, the programmeddevice120 requires the user or video submitter to input at least one descriptor or a minimum amount of descriptors through the primaryvideo categorizer interface287. If the video submitter fails to do so, the programmeddevice120 blocks, prevents or disables the distribution of the applicable compilation video to the home interface54 (FIG. 3A). Accordingly, such video will not be published through thehome interface54.
In another embodiment, the programmeddevice120 requires the user or video submitter to input a minimum amount of descriptors through the primaryvideo categorizer interface287 and the secondaryvideo categorizer interface288. If the video submitter fails to do so, the programmeddevice120 blocks, prevents or disables the distribution of the applicable compilation video to the home interface54 (FIG. 3A). Accordingly, such video will not be published through thehome interface54.
Referring again toFIG. 40B, in response to the user's selection of thenext element291, the programmeddevice120 generates apublic publication interface290 in accordance with thepublication module31 as illustrated byFIG. 40C. As shown, thepublic publication interface290 shows thefirst frame292 of thecompilation video280. Also, the public publication interface290 displays a plurality of data fields, including: (a) a caption field enabling the user to enter text describing the video, such as “Power Bornfreedom's Triple-Double!;” (b) a game date field; (c) an athlete field for the name of the highlighted athlete who is registered with the system13, which is selectable from a list of athletes via a search interface; (d) a video shooter field for the name of the videographer or video producer (e.g., “MadSkilz TV”) registered with the system13 who is selectable from a list of video producers via a search interface; (e) a home field enabling the user to enter text describing the name of the home team, such as “Brightmore High School,” which may be selectable via a search interface; (f) a mascot field for the name of the home team's mascot, which may be pre-populated based on the selection of the home team; (g) a visitor field enabling the user to enter text describing the name of the visitor team, such as “Calvary High School,” which may be selectable via a search interface; (h) a league field for entry of the applicable sports league (e.g., “Chicago Public League”) which may be selectable via a search interface; and (i) a plurality of statistics fields, such as a points field, steals field, assists field, blocks field, rebounds field and turnovers field. If the user inputs statistics during the recording sessions, as described above, the programmeddevice120 automatically pre-populates the statistics fields with the different totals of the statistics input by the user. For example, thepublic publication interface290 may automatically display “18” in the points field, “12” in the assists field, “10” in the rebounds field, “3” in the blocks field, and “5” in the steals field. If any of the statics fields are blank because the user decided not to record or input the applicable statistic during the recording session, the user can manually enter statistical text in such field. Also, the user can override any of the pre-populated statistics fields by changing the statistical text in such field.
Thepublic publication interface290 also displays a sound field or sound symbol. By selecting the sound symbol, the user can upload, download or otherwise capture a desired sound track or musical recording. Depending upon the embodiment, the source of the sound track can be the local data storage of the programmeddevice120 or a web server. In an embodiment, once the user captures the sound track, the programmeddevice120 automatically: (a) cuts or trims the length of the sound track to match the length of thecompilation video280; and (b) incorporates the sound track into thecompilation video280, replacing the original audio of thecompilation video280 with the sound track.
After these steps, the user can press thepublic post element294. In response, the programmeddevice120 generates thefront video interface296 as illustrated inFIG. 41. In an embodiment, thefront video interface296 includes: (a) at least oneadvertisement section298 providing space for a promotion or advertisement of a company or organization, such as thesports drink advertisement300; and (b) anathlete portrait section302 providing space for an image or photo of the athlete displayed in theapplicable compilation video280, such as theathlete photo304; (c) avideo summary section306 displaying the key information regarding the athlete, the event and the athlete's statistics, such as the athlete's name (e.g., Power Bornfreedom), jersey number (e.g., #15), high school (e.g., Brightmore High School), the date (e.g., Nov. 8, 2018), the final score of the game (e.g., Brightmore: 74, Calvary: 64), and the athlete's points, assists, rebounds, blocks and steals.
In an embodiment, the participant center interface308 (FIG. 51) enables the user (e.g., the athlete or the athlete's friend or parent) to capture and store a photo of the athlete, such as the athlete photo310 shown inFIG. 41. In such case, the programmeddevice120 automatically loads and displays the athlete photo310 in theathlete portrait section302. If there is no prestored athlete photo, thefront video interface296 enables the user to take a photo of the athlete or upload or download the athlete's photo from the programmeddevice120 or a webserver. Then, thefront video interface296 enables the user to capture and display such photo in theathlete portrait section302. If the user adds no photo to theathlete portrait section302, the programmeddevice120 adds the first frame of thecompilation video280 to theathlete portrait section302.
In publishing thecompilation video280, the programmeddevice120 transfers the thecompilation video280 to the one or more data storage devices12 (FIG. 1). Using the search interface312 (FIG. 3A), users (e.g., participants, fans and other non-participants) can locate, access and view thecompilation video280, such as thecompilation videos60,61,62 shown inFIG. 3A.
When the user clicks or selects a compilation video, such as the compilation video60 (FIG. 3A), the programmeddevice120 displays thesocial interface314 as illustrated inFIG. 42A. In an embodiment illustrated inFIG. 42A, thesocial interface314 includes: (a) thefront video interface296, which functions as the introductory frame or introductory image of thecompilation video280; (b) the name, trademark oridentifier316 of the video shooter, for example, “MadSkilz TV”; (c) aflame quantity318; (d) aview quantity320; (e) ashare element322, the selection of which enables users to share thecompilation video60 with, or send thecompilation video60 to, other users; and (f) acomment element324, the selection of which enables users to postcomments325 related to thecompilation video60.
When the user taps, pauses or finishes watching thecompilation video60, the programmeddevice120 displays aflame rating interface326 as illustrated inFIG. 42B. Theflame rating interface326 includes: (a) asmall flame326 associated with a count of one flame, a relatively low level of likeness; (b) amedium flame327 associated with a count of two flames, a moderate level of likeness; and (c) alarge flame331 associated with a count of three flames, a relatively high level of likeness. Thesystem13 keeps count of the quantity of flames input by users, and thesystem13 displays the current flame total at theflame quantity318.
In an embodiment, thesystem13 calculates a fire rating390 (FIG. 52A), an internal metric, that depends on the current quantity of flames and the current quantity of views. In an embodiment, the fire rating is equal to the current quantity of flames divided by the current quantity of views resulting in a flames per view metric. This ratio reflects the assumption that a highly interesting video should have a relatively high quantity of flames per view.
In an embodiment, thesystem13 includes a video auto-deletion function to automatically purge the one or moredata storage devices12 of redundant videos—videos that highlight the same athlete in the same event. This video auto-deletion function reduces clutter and saves storage space in the one or moredata storage devices12. Also, this video auto-deletion function simplifies the home interface54 (FIG. 3A) so that users do not have to sort through redundant videos. In an embodiment, thesystem13 determines the first-in time at which eachcompilation video280 is published (e.g., 10:20 pm Eastern Time, Nov. 26, 2018), and thesystem13 also determines a video profile associated with such video, such as the name of the highlighted athlete, the date of the game, and the names of the home and visitor teams. Thesystem13 has a setting for a designated time window. The time window starts or opens at the first-in time, and the time window ends or closes at a designated time point following such first-in time (e.g., four hours after the first-in time or 2:20 am Eastern Time, Nov. 27, 2018). Thesystem13 determines the fire rating (e.g., flames per view) of eachsubsequent compilation video280 with the same video profile that is published within the time window. Thesystem13 compares the fire ratings and determines which one ofsuch compilation videos280 has the highest fire rating. Next, thesystem13 automatically deletes all of theother compilation videos280. At that point, only thecompilation video280 with the highest fire rating, considered the winning video, remains stored in the one or moredata storage devices12.
In an embodiment, thesystem13 automatically blocks the publication ofcompilation videos280 of such video profile once the time window ends or closes. In this case, the programmeddevice120 automatically displays a closed indicator (e.g., “POSTING TIME ENDED” or “CLOSED”) when the user enters enough data in the public publication interface290 (FIG. 40C). For example, the user may enter the game date, athlete name, home team and visitor team. In response, the programmeddevice120 may display “CLOSED” and disable the submitelement294.
In an embodiment, thesystem13 enables the athlete highlighted in the winningcompilation video280 to replacesuch compilation video280 with analternate compilation video280 published by the athlete. This may be desirable, for example, if such athlete is displeased with the quality of the winningcompilation video280. Depending upon the embodiment, thesystem13 can also enable such athlete to takedown or delete winningcompilation videos280 that emphasize such player's mistakes or poor or unflattering performance.
In the example illustrated inFIG. 43A, the user selected Athlete Lowlights in the secondaryvideo categorizer interface288. The Athlete Lowlights category is associated with a private setting corresponding to theprivate posting interface328. In response to the user's input through theprivate post element329, the programmeddevice120 transfers thelowlight compilation video280 to the participant module32 (FIG. 1). This makes thelowlight compilation video280 privately accessible to the user through theparticipant center interface308 shown inFIG. 50A, as described below.
In many cases, relatively low profile events, such as amateur sports games and high school games, receive little, if any, media coverage. Many of the events are not broadcast live by news channels. As a result, the participants do not receive timely exposure from the events, which can result in lost opportunities. Furthermore, the information that does circulate, such as a player's statistics or performance, can be inaccurate. For example, a high school team may have a game that is not covered by the local news media. When the game finishes at 9:00 pm on a Friday, a spectator might publish a Twitter™ message, such as “Chris Carlson scores 34 in Brightmore Tigers' win over Glendale Bears!” In this example, such information is false or fake news. The truth is that Chris Carlson scored 22 points, and the Glendale Bears won the game. The misinformation can be inaccurate or misleading. This can cause harm to the reputation and opportunities of the event participants.
In an embodiment, the verification module34 (FIG. 1) in conjunction with thepublication module31, described above, provides an improvement to overcome or lessen these disadvantages. In an embodiment, theverification module34 enables a crowd or relatively large pool of users to help verify or increase the reliability of the event information provided by submitters ofcompilation videos280.
As described above, the public publication interface290 (FIG. 40C) includes a plurality of data fields related to the event (e.g., game). Any user attending the game can use any programmeddevice120 to enter text into these fields and press the submit element294 (FIG. 40C). Thesystem13 processes the event data entered by each such user.
In an embodiment, theverification module34 includes verification logic that is executable to compare the event data provided by one user for a certain video profile to the event data provided by the other users for the same video profile. If thesystem13 determines that the event data of a designated quantity of users match, thesystem13 confirms such event data as verified and indicates the verification by displaying a verification indicator330 (FIG. 42A) within thesocial interface314.
For example, thirty users may submit thirtycompilation videos280 with the same video profile within one hour after the end of a Friday night high school basketball game, resulting in a sequence of event data submissions one through thirty as follows:
|
| Submission | Final Score | Comparison | |
|
| 1 | Brightmore 74,Calvary 68 | Match |
| 2 | Brightmore 74,Calvary 68 | Match |
| 3 | Brightmore 70, Calvary 66 |
| 4 | Brightmore 72, Calvary 85 |
| 5 | Brightmore 74,Calvary 68 | Match |
| 6 | Brightmore 74,Calvary 68 | Match |
| 7 | Brightmore 74,Calvary 68 | Match |
| • | • | • |
| • | • | • |
| • | • | • |
| 30 | • | • |
|
In this example, thesystem13 includes a verification factor that requires a minimum of five final score submissions to match each other. Once the first five submissions have matching final scores, thesystem13 designates the final score as verified or confirmed. Then, thesystem13 automatically either: (a) adds the confirmed event data316 (FIG. 41) to thefront video interface296 of each one of thecompilation videos280; or (b) changes the existing, original data ofsuch compilation videos280 to match the confirmedevent data316. This verification or confirmation functionality increases the credibility and objectivity of the video information published through thesystem13, which enables recruiters, colleges and other users to place greater reliance on the video information for athlete evaluation purposes.
In another embodiment illustrated inFIGS. 44-48, thesystem13 includes an empirical evidence-based verification or confirmation system. As indicated bystep332, the programmeddevice120 receives a video submission from a user incorporating a report or event data that includes text of the home team's name, home team's mascot, visitor team's name, home team's final score, and visitor team's final score. As indicated bystep334, based on the user's permission, thesystem13 tracks the geographic location of the programmeddevice120 upon receiving the report or within a relatively short time period (e.g., five seconds) after receiving the report. In an embodiment, thesystem13 is operatively coupled to a webserver having the addresses of the home team. Based on that address information and the location tracking, thesystem13 determines whether the current location of the programmeddevice120 is within a designated area surrounding (or radius from) the venue of the home team as indicated bydecision diamond336. For example, thesystem13 may determine whether the programmeddevice120 is within one thousand feet or one-half mile from the stadium of the home team. If the answer is no, the programmeddevice120 indicates that the confirmation or verification is incomplete as indicated bystep338 and verification failure indicator339 (FIG. 48C). This is based on the reasoning that the report is more likely to be accurate if it is received by a user who is physically present at or nearby the location of the event. If the answer is yes, the programmeddevice120 generates an image submitted by the user pertaining to the event as indicated byblock340. In an embodiment, the image includes a photo of an outcome indicator342 (FIG. 46), such as the physical scoreboard mounted to the stadium wall or otherwise coupled to the stadium or gymnasium. Next, thesystem13 receives and converts the image evidence to text and analyzes the text, determining the following information displayed on the outcome indicator342: the home team's name, home team's mascot's name, visitor team's name, home team's score, and the remaining game time as indicated byblock344. Thesystem13 can convert such image to text through optical character recognition (OCR) or any other suitable conversion method.
Next, as indicated bydecision block346, thesystem13 determines whether the text extracted from theoutcome indicator342 indicates: (a) zero seconds of remaininggame time347; and (b) ahome score348 andvisitor score350 that match the corresponding data reported with thecompilation video280 submitted by the user. If the answer is no, the programmeddevice120 indicates that the confirmation or verification is incomplete as indicated bystep338 and verification failure indicator339 (FIG. 48C). If the answer is yes, thesystem13 determines, as indicated bydecision block352, whether thesystem13 has received X number of one or more reports of the same video profile that: (a) have no discrepancy with a certain percentage of the other reports; and/or (c) have no discrepancy with the text evidence extracted from theoutcome indicator342. Next, as indicated byblock354, thesystem13 filters the data reported with thecompilation video280, determines any such data that conflicts with the text evidence extracted from theoutcome indicator342, and automatically replaces such data with the applicable text data derived from theoutcome indicator342. The programmeddevice120 then generates the verification success indicator355 (FIG. 48B) and the verification indicator330 (FIG. 42A). As indicated byblock356, thesystem13 then transfers the verified data to theparticipant module32 of the athlete who is identified within the video profile ofsuch compilation video280. Next, as indicated bystep358, the programmeddevice120 indicates benefits to such athlete based on such verified data, as described below.
In another embodiment illustrated inFIG. 45, the programmeddevice120 receives a video submission from a user incorporating a report or event data as indicated byblock361. The report or event data can include text of the home team's name, home team's mascot, visitor team's name, home team's final score, and visitor team's final score. The programmeddevice120 then generates one or more images submitted by the user pertaining to the event as indicated byblock363. In an embodiment, the one or more images include a photo363 (FIG. 47A) of the outcome indicator342 (FIG. 46) and a photo365 (FIG. 47B) of a mascot name364 (FIG. 46) painted or mounted to the stadium wall or otherwise coupled to the stadium or gymnasium.
Themascot name364 can be indicated on a banner, on a painted section of a wall, on theoutcome indicator342 or on another physical display medium366 (FIG. 46). In the example shown, the mascot name is “TIGERS.” Next, as indicated bydecision diamond365, thesystem13 determines whether the photo of themascot name364 was submitted by the user (and received by the system13) within a designated period of time (e.g., five seconds) after thesystem13 received the user's submission of the photo of theoutcome indicator342. If the answer is no, the programmeddevice120 indicates that the verification is incomplete as indicated byblock367 and verification failure indicator339 (FIG. 48C). This is based on the reasoning that, if the user is actually at the site of the game, the user will be able to photograph theoutcome indicator342 and themascot name364 in quick succession.
In the example illustrated inFIG. 47, the programmeddevice120 displays image capture interfaces369,371. Theimage capture interface369 enables the user to photograph and upload thescoreboard photo363, and theimage capture interface371 enables the user to photograph and upload themascot banner photo365.
Referring back toFIG. 45, if the answer todecision diamond365 is yes, thesystem13 receives and converts the image evidence to text and analyzes the text, determining the following information displayed on the outcome indicator342: the home team's name, home team's mascot's name, visitor team's name, home team's score, and the remaining game time as indicated byblock369. Thesystem13 can convert such image to text through OCR or any other suitable conversion method.
Next, as indicated bydecision block373, thesystem13 determines whether the text extracted from theoutcome indicator342 indicates: (a) zero seconds of remaininggame time347; and (b) ahome score348 andvisitor score350 that match the corresponding data reported with thecompilation video280 submitted by the user. If the answer is no, the programmeddevice120 indicates that the confirmation or verification is incomplete as indicated bystep367 and verification failure indicator339 (FIG. 48C). If the answer is yes, thesystem13 determines, as indicated bydecision block375, whether thesystem13 has received X number of one or more reports of the same video profile that: (a) have no discrepancy with a certain percentage of the other reports; and/or (c) have no discrepancy with the text evidence extracted from theoutcome indicator342. Next, as indicated byblock377, thesystem13 filters the data reported with thecompilation video280, determines any such data that conflicts with the text evidence extracted from theoutcome indicator342, and automatically replaces such data with the applicable text data derived from theoutcome indicator342. The programmeddevice120 then generates the verification success indicator355 (FIG. 48B) and the verification indicator330 (FIG. 42). As indicated byblock379, thesystem13 then transfers the verified data to theparticipant module32 of the athlete who is identified within the video profile ofsuch compilation video280. Next, as indicated bystep381, the programmeddevice120 indicates benefits to such athlete based on such verified data, as described below.
As illustrated inFIGS. 48A-48C, the programmeddevice120 displays: (a) the verification in process indicator382 (e.g., an image or animation of a basketball moving toward a hoop) during the verification processes described above; (b) the verification success indicator355 (e.g., an image or animation of a basketball within a hoop) in response to a successful verification of reported video data; and (c) a verification failure indicator339 (e.g., an image or animation of a basketball outside of a hoop) in response to a failure of an attempted verification described above.
Many types of events, such as sports games, talent shows, theatrical plays and concerts, have venues where relatively large numbers of people attend. At the end of the event, the participants, their friends in the audience and other attendees often are hungry and wish to visit a local restaurant or eatery for a meal or snack. The food providers or restaurants compete with each other for these customers. Often times, restaurants located further from the venue receive less customers from the event than those restaurants located closer to the venue.
At or after the end of the event, thesystem13 receives, verifies and transfers the event outcome data to theparticipant module32 as described above. Referring toFIGS. 49A-49B, in an embodiment, thesystem13 determines when a logged-in user is a participant (e.g., an athlete) who is registered with thesystem13, as described below. For example, a registered athlete may access thesystem13 through a programmeddevice120 in the locker room shortly after the game ends. If the athlete's team won the game, the programmeddevice120 displays awinner benefit interface341 as illustrated inFIG. 49A. If the athlete's team lost the game, the programmeddevice120 displays aloser benefit interface343 as illustrated inFIG. 49B.
Thewinner benefit interface341 displays: (a) the verifiedevent outcome data344; (b) awin indicator349, such as “Enjoy a treat for your win!”; (c) anexpiration notice348, such as “Expires at 11:37 pm”; (d) a plurality of award indicators orbenefit indicators350, such as free food items offered by various fast food restaurants; and (e)benefit terms352, such as “Good for you and4 friends!”
Theloser benefit interface343 displays: (a) the verifiedevent outcome data344; (b) awin indicator354, such as “Enjoy a treat for your effort!”; (c) anexpiration notice348, such as “Expires at 11:37 pm”; (d) a plurality of award indicators orbenefit indicators356, such as food discounts and free food items offered by various fast food restaurants; and (e)benefit terms358, such as “Good for you and 2 friends!” In this example, the value of thebenefit indicators356 is less than the value of thebenefit indicators350. Similarly, thebenefit terms358 are less favorable than thebenefit terms352. It should be appreciated that, in other examples, theinterfaces341,342 can have different expiration notices348 and other differences that grant more favor to the winning registered player than the losing registered player.
With the benefits indicated by thewinner benefit indicator340 or the losingbenefit indicator342, as applicable, the registered athlete can visit the applicable restaurant, before the applicable expiration time, with companions or friends. Upon arrival, for example, a winning athlete can obtain five items of large fries for the athlete and four friends. The transaction can be performed through different methods. In an embodiment, the programmeddevice120 displays a unique code, such as a unique numeric or alphanumeric code or a scannable code (e.g., a 1D or 2D barcode, such as QR code datamatrix). In another embodiment, the programmeddevice120 generates a signal, such as a radio frequency (“RF”) or infrared radiation (“IR”) signal. In yet another embodiment, to enroll for the benefit indicated atwinner benefit indicator341 and the losingbenefit indicator342, the benefit providers or restaurants require the participants to create loyalty card accounts with the restaurants, associating the participants' phone numbers with their accounts. Depending upon the embodiment, the cashiers of the restaurants can ascertain the benefits awarded to the participants by: (a) entering codes provided by the participants; (b) scanning barcodes displayed on the participants' programmeddevices120; (c) establishing an electronic communication between the point of sale machines and the programmeddevices120 to receive signals from the programmeddevices120; (d) entering the participants' phone numbers; or (e) any other suitable benefit transfer method. In an embodiment, each benefit provider (e.g., restaurant) has a webserver, database or benefit source44 (FIG. 1) that is operatively coupled to thesystem13 through thenetwork16. Such benefit provider manages the distribution and accounting of benefits (e.g., discounts and freebies) to each unique event participant who is registered through thesystem13.
In an embodiment, the programmeddevices120 are enabled for near-field communication (“NFC”). For example, the programmeddevices120 can have RF transceivers, NFC protocols and NFC code operable to perform NFC with the point of sale devices of restaurants and other providers. For example, the NFC code can include a mobile wallet app such as Google Wallet™ or Apple Pay™. In another embodiment, the participant module32 (FIG. 1) includes computer code the enables users to load their credit, debit, gift and loyalty cards to thesystem13 so that they may use theirprogrammed devices120 to make payments and perform transactions in stores. In yet another embodiment, thesystem13 is operatively coupled to the Samsung Pay™ platform to enable such functionality.
As described above, the user can tap or activate themenu element81 to cause the programmeddevice120 to display the features interface82 (FIG. 3B). At any time, the user can tap or activate theparticipant center element90 of thefeatures interface82. In response, the programmeddevice120 will display theparticipant center interface308, as illustrated inFIG. 50A. Theparticipant center interface308 has: (a) apublic zone360 that archives and stores the registered participant's information, images and videos intended for public viewing; and (b) aprivate zone362 that archives and stores the registered participant's information, images and videos that are intended to be kept private. In the example shown, thepublic zone360 includes personal data, highlight compilation videos, one or more interview videos for viewing by colleges and recruiters, one or more reference videos provided by teachers or coaches, a personal photo of the participant, a biography page regarding the participant, and a video distribution element for sending desired ones of these videos to colleges, recruiters or others. In the example shown, theprivate zone362 includes lowlight videos, development videos (e.g., videos of the participant's training sessions) and a list of the participant's gift cards and sponsors.
Thesystem13 publishes thepublic zone360 to the public, and thesystem13 blocks public access to theprivate zone362. The programmeddevice120 enables the participant to provide select people (e.g., trainers, coaches, family members or recruiters) with access to theprivate zone362. It should be understood that the video generator28 (FIG. 1) could have been used to record and capture all of the videos within thepublic zone360 and theprivate zone362.
As illustrated inFIG. 50B, the programmeddevice120 displays apersonal data interface383 in response to the participant's activation of thepersonal data element366. In the example shown, thepersonal data interface383 has a plurality data fields for collectingpersonal data368. In the example shown, thepersonal data368 includes the participant's name, zip code, birthdate, school, GPA, ACT score, SAT score, sport, coach's name, position, height, and weight.
As illustrated inFIG. 1, thesystem13 enables the participant to setup data feeds from a plurality of data sources18 (e.g., webservers or databases) of entities including, but not limited to, schools38,healthcare providers40, andtesting organizations42. In an embodiment illustrated inFIG. 51A, the programmeddevice120 displays a personaldata verification interface370. Thesystem13, through communication with the data sources18, automatically checks for matches between thepersonal data368 input by the participant and the corresponding data documented in the records of the data sources18. If there is a match, the personaldata verification interface370 indicates the match as a verification. In the example shown, the verifications are indicated by checkmarks. In addition, the programmeddevice120 display averification progress interface372 that indicates the participant's progress in obtaining verifications. In the example shown, theverification progress interface372 displays aprogress meter374.
In response to the participant's activation of the highlight video element378 (FIG. 50A), the highlight video interface376 (FIG. 52A) displays thehighlight compilation videos380,382,384 generated by thevideo generator28. Also, thehighlight video interface376 displays afire rating meter386. Thefire rating meter386 displays the fire rating390 (as described above in flames per view) of the participant's highest ratedvideo380.
As illustrated inFIG. 52B, the programmeddevice120 displays aninterview video interface392 in response to the participant's activation of the interview video element394 (FIG. 50A). Theinterview video interface392 displays the participant'sinterview video396.
As illustrated inFIG. 53A, the programmeddevice120 displays areference video interface398 in response to the participant's activation of the reference video element400 (FIG. 50A). Thereference video interface398 displays the participant'sinterview videos402,404, together with text regarding the interview videos. In the example shown, the text states the name and title of the interviewee, together with the date of the interview.
As illustrated inFIG. 53B, the programmeddevice120 displays abiography interface406 in response to the participant's activation of the biography page element408 (FIG. 50A). Thebiography interface406 displays a plurality of personal data fields410. The participant can enter his/her data in the personal data fields410.
Referring back toFIG. 50A, the participant can press or select the send videos element409 of thepublic zone360. In response to such selection, the programmeddevice120 displays asend videos interface411, as illustrated inFIG. 54A. In the example shown, thesend videos interface411 displays the first frames of thehighlight videos380,382,384, and the participant selected thehighlight compilation video380. In response to such selection, the programmeddevice120 displays arecipient interface413. Therecipient interface413 displays a plurality of selectable recipients, which, in the example shown, include a Facebook™ account, an email account linked to a list of recruiters, a Twitter™ account, and a plurality of email addresses of designated contacts of a plurality of colleges A, B and C. Therecipient interface413 also displays asearch field415 that enables the user to enter text to search for a prestored recipient. In response to the participant's selection of one or more of therecipient elements417, the programmeddevice120 emails, sends or otherwise transfers the selectedhighlight compilation video380 to the recipients associated with the selectedrecipient elements417.
In response to the participant's activation of the lowlight video element414 (FIG. 50A) in theprivate zone362, the programmeddevice120 displays alowlight video interface412 as illustrated inFIG. 55A. Thelowlight video interface412 displays thelowlight compilation videos416,418 generated by thevideo generator28. Also, thelowlight video interface412 displays text associated with thelowlight compilation videos416,418, such as “Weak defense, Nov. 12, 2017” or “Sloppy passing; not boxing-out, Dec. 8, 2017.”
As illustrated inFIG. 55B, the programmeddevice120 displays adevelopment video interface420 in response to the participant's activation of the development video element422 (FIG. 50A). Thedevelopment video interface420 displays thedevelopment compilation videos424,426 generated by thevideo generator28. Also, thedevelopment video interface420 displays text associated with thedevelopment compilation videos424,426, such as “63 of 100 threes, Apr. 13, 2018” or “Two-hand dunk, Oct. 5, 2017.”
As illustrated inFIG. 56, the programmeddevice120 displays agift card interface428 in response to the participant's activation of the gift card element430 (FIG. 50A). Thegift card interface428 displays a list of the gift card accounts432 of the various service providers and merchants with whom the participant is registered. As shown, thegift card interface428 displays the purse values of the gift card accounts432.
As illustrated inFIG. 57A, the programmeddevice120 displays asponsor level interface434 in response to the participant's activation of the sponsor element436 (FIG. 50A). In the example shown, thesponsor level interface434 displays: (a) an athlete rating438 that is limited to or is derived from one or more of the following factors: the participant's athletic performance statistics, the flame per view rating390 (FIG. 52A), the participant's biographical data, or any suitable combination thereof; (b) a student rating440 that is limited to or is derived from one or more of the following factors: the participant's school grades, ACT score, SAT score or any suitable combination thereof; and (c) thefollower count442 for the followers of the participant. Based on one or more of the athlete rating438, the student rating440, and thefollower count442, thesystem13 determines the sponsor level of the participant. In the example shown, thesponsor level interface434 displays asponsor meter444 having a plurality of thresholds indicated by $, $$ and $$$. In this example, the participant's sponsor level has risen to the $$ sponsor level.
In response to the get sponsoredelement446, the programmeddevice120 displays the sponsors interface448 as illustrated inFIG. 57B. The sponsors interface448 displays the list of participatingsponsors450. In the example shown, thesponsors450 include sports shoe manufacturers and sports drink manufacturers. Thesponsors450 have certain terms and conditions regarding the sponsorship. Once the participant decides upon one or more of thesponsors450, the participant can proceed with one or more of the sponsorships offered to the participant. In the example shown, the participant selected theAdidas element452 corresponding to the sponsorship offered by the Adidas™ company. In response, the programmeddevice120 displays thesponsor account interface454 as illustrated inFIG. 57C. In the example shown, thesponsor account interface454 displays information regarding the Adidas™ sponsorship, including the sponsor's name, the expiration date of the sponsorship, the sponsorship level, the purse or wallet value of the sponsorship, the gift awarded, and the grant of free academic, test preparation courses. In the example shown, the participant will receive $239.17 in spending money, a pair of free Adidas™ basketball shoes and a free ACT/SAT preparation course.
It can be difficult for event participants to find suitable partners or assistants for the pursuit of their objectives. For example, it can be challenging for athletes to find suitable AAU teams, sports camps, college recruiters, trainers and other partners. The connector module36 (FIG. 1) provides an improvement to help overcome this challenge. Referring toFIG. 58A, the programmeddevice120 executes theconnector module36 to display aconnector interface456 in response to the user's selection of theconnection symbol80 or the connector element92 (FIG. 3B).
Theconnector interface456, shown inFIG. 58A, enables the user (e.g., an athlete, other participant or parent of a participant) to search for, review, assess and matchup with providers of services, products or opportunities, such as people, organizations or businesses. Theconnector interface456 displays alisting element458 and aconnection facilitator element460. In response to the user's selection of thelisting element458, the programmeddevice120 displays alisting interface462 as illustrated inFIG. 58B. Thelisting interface462 is usable by users who are providers, such as owners, operators, employees, agents or representatives of businesses or organizations, including, but not limited to, AAU teams/clubs, hosts of sports camps, athletic programs, training businesses, recruiting businesses, physical therapy businesses, healthcare providers and other providers of services or goods. As shown, thelisting interface462 displays a plurality of data fields, including, but not limited to, category (e.g., trainer or AAU team), name, address, description, logo, tryout schedule and requirements, practice schedule, game schedule, fees, director's name, website address, contact information, payment method and other information.
In response to the user's selection of theconnection facilitator element460, the programmeddevice120 displays aconnection search interface464 as illustrated inFIG. 59A. Theconnection search interface464 displays atype filter466, alocation filter468 and asort element470. The activation of thetype filter466 enables the user to select a desired category or type of provider from a list of types or categories of providers. In the example shown, the list includes AAU team, camp, college recruiter, physical therapist, sports agent, trainer and tutor. Thelocation filter468 enables the user to filter the service/goods provider by specified location. The programmeddevice120 displays the search results based on the sort preferences set by the user through thesort element470.
In the example shown inFIG. 59A, the user selected theAAU team category472 for the category ortype466, enteredzip code60649 for thelocation468, and selected rating474 for thesort element470. In response, the programmeddevice120 displayed the search results interface476. In this example, the search results interface476 displays a list of AAU basketball clubs, including the quantity of reviews and star rating on a scale of one to five stars. The club with the highest rating is displayed at the top of the list.
Continuing with this example, the user selected theChicago Blaze club478. In response, the programmeddevice120 displayed theprovider interface480 as illustrated inFIG. 60A. Theprovider interface480 displayed a plurality ofreview interfaces482,484,486. Each of the review interfaces482,484,486 is associated with a compilation video or other video produced by a user through thevideo generator28 as described above. In this embodiment, eachreview interface482,484,486 displays a locked mode by default as follows: (a) avideo area488 that is blank or otherwise masks the applicable video; (b) astar rating490; (c) areview date492; and (d) atext area494 that is bank or masks the text of the applicable review. To unlock the reviews, the user can select a service plan from a plurality of different service plans497 displayed by thereview unlock interface499 as illustrated inFIG. 60B. The user can then pay for and purchase a selected one of the plans by selecting thepurchase element498. After the user makes the payment, the programmeddevice120 transitions to the unlocked mode.
In the example shown inFIG. 61A, the programmeddevice120 unmasked the reviews and videos within the review interfaces482,484,486 (FIG. 60A). For example, thereview interface482 states, “By Jane Doe on Aug. 25, 2017. Watch this coach screaming at 8th graders. This team is bad news.” Thereview interface482 also includes acompilation video496 produced by Jane Doe. Thecompilation video496 shows the coach exhibiting the screaming behavior during a practice or game of the Chicago Blaze.
For parents of participants under the age of eighteen, it can be difficult to research and identify suitable organizations for their children. For example, most parents of student athletes rely on word-of-mouth information regarding AAU teams. This is because there is little online information regarding many of these teams, and there is no readily-accessible, reliable resource that provides transparency into the team activities or otherwise facilitates the integrity, accuracy and objectivity of the information. Consequently, parents often mistakenly select AAU teams that are lead or coached by adults who are lacking in ethics and competence or who engage in nepotism. This exposes children and youth to hostile environments involving bullying by coaches, embarrassment or ridicule by coaches, poor role models of coaches engaged in fighting, profanity and confrontations with referees and others, physical and psychological abuse by coaches and other acts that are harmful to the self esteem and development of children and youth. The provider interface480 (FIG. 60A) provides an improvement to help overcome this problem. For example, theprovider interface480 enables parents to see inside an organization (e.g., AAU team) by watching truthful, review-based videos generated through thevideo generator28 as described above.
If the user is interested in matching-up with, contracting with, joining or otherwise connecting with a provider who is listed through listing element458 (FIG. 58A), the user can select the provider's name. In the example shown inFIG. 61A, the user selected theChicago Blaze name498, and, in response, the programmeddevice120 displayed theprovider profile500 regarding the Chicago Blaze club as illustrated inFIG. 61B. Theprovider profile500 includes a list of hyperlinks to detailed information regarding the Chicago Blaze club as well as a plurality of selectable options. In this example, the user selected thegirls option502 and thepayment element504. Thepayment element504 enables the user to submit an electronic payment to join the Chicago Blaze club.
Conventionally, many providers such as AAU clubs, are not equipped to accept credit card or electronic payments. They require cash payments. The lack of receipts and handling of cash can cause security and fraud risks for payers. In an embodiment, the user can make one-time payments and periodic payments to the listed providers through theprovider profile500. This provides an improvement in security and convenience for athletes, participants and parents.
In an embodiment, the programmeddevice120 is operable to display anitem order interface506 as illustrated inFIG. 62A. In the illustrated embodiment, the purchasable item includes a wearable device, abracelet508 as illustrated inFIG. 62B. Thebracelet508 includes anelectrical element510. Theorder interface506 enables the user to customize thebracelet508 with the user's name, a desired slogan, expression or quote, and the desired color. By selecting thepayment element512, the user can pay for and order thebracelet508.
In an embodiment, the programmeddevice120 is operable to display anitem order interface514 as illustrated inFIG. 63A. In the illustrated embodiment, the purchasable item includes a wearable device, ashoestring tag516 as illustrated inFIGS. 63B and 63C. Theshoestring tag516 includes anelectrical element510. Theorder interface514 enables the user to customize theshoestring tag516 with the user's name (e.g., “J. SMITH”), an identification or member ID number (e.g., “#2849”) generated by thesystem13, a desired slogan, expression or quote (e.g., “NEVER QUIT”), and the desired color. By selecting thepayment element520, the user can pay for and order theshoestring tag516.
In this embodiment, theshoestring tag516 includes abody522 that defines a plurality of fasteners or couplers which, in the example shown, includestring receiving holes524,526. Thebody522 has a downwardly-curved, arc shape as shown. It should be appreciated, however, that thebody522 can be flat, wavy or have any other suitable shape. As shown inFIG. 64A, thestring receiving holes524,526 are configured to receivesegments528,530, respectively, of ashoestring536 of ashoe534. Theshoestring tag516 is removably coupled to theshoestring536 which, in turn, is removably coupled to theshoe534.
In an embodiment, theelectrical element510 includes: (a) an antenna, transmitter or radiator operable to generate a wireless signal, such as a suitable RF; (b) a receiver operable to receive such a wireless signal; (c) a transceiver operable to generate and receive such a wireless signal; (d) a sensor operable to monitor or detect events and conditions related to the user who is wearing thebracelet508 orshoestring tag516 or the environment in which the user is running, walking, standing or participating; or (e) a memory unit operable to store data. In an embodiment, theelectrical element510 includes any suitable combination of the foregoing components. In an embodiment, the sensor has circuitry, including a data processor and memory, configured to sense foot speed, acceleration, impact, stress, fastest speed, the heights of jumps, biometric activity of the wearer and other performance-related factors that occur throughout the game or event.
In an embodiment, theelectrical element510 has circuitry coupled to a miniature battery power source. In another embodiment, theelectrical element510 includes a passive radio-frequency identification (“RFID”) module having a circuit configured to: (a) store and process information that modulates and demodulates external RF signals; (b) a power receiver operable to receive electrical power from the external RF signals; and (c) a transceiver operable to receive and transmit the RF signals.
Theelectrical element510 is configured to communicate with or transmit signals to one or more external transceivers. Depending upon the embodiment, the external transceivers can be components of one or moreprogrammed devices120 or components of one or more sensors installed in the facility where the wearer is performing. In an embodiment, each external transceiver includes an RF transceiver operable to send high frequency RF signals to, and receive high frequency RF signals from, theelectrical element510.
In operation of an example, an athlete installs theshoestring tag516 on the athlete'sshoe534 as illustrated inFIG. 64. Theshoestring tag516 is operable to receive and respond to a signal generated by an external RF transceiver, such as aprogrammed device120 that is paired with theshoestring tag516. A member of the audience, such as a parent of the athlete, is seated in bleachers holding the programmeddevice120. The programmeddevice120 wirelessly communicates with theshoestring tag516. Theelectrical element510 senses and stores information regarding the athlete's performance throughout the game. The programmeddevice120 communicates with theshoestring tag516 to receive such information. For example, as illustrated inFIG. 65, the programmeddevice120 generates the athlete metrics interface538. Based on the information received from theshoestring tag516, the athlete metrics interface538 displays data, including: the peak acceleration or history of accelerations; peak speed or history of speeds; peak vertical jumping height or history of jumping heights; playing time or hours trained; steps taken; and distance from the programmeddevice120 to theshoestring tag516.
In another embodiment, theelectrical element510 is configured to generate an energy signature, such as an RF signature, infrared light or other light within the invisible spectrum. In this embodiment, the programmeddevice120 has a thermal imaging device, infrared radiation reader, video camera or other sensor that is configured to continuously track and detect the energy signature. Using the energy signature, the video generator28 (FIG. 1) generates a tracking image on or adjacent to the video-recorded image of the participant in the event. In the example shown inFIG. 66, thevideo generator28 generates the trackingimages540,542 under the athlete's feet. It should be appreciated that the trackingimages540,542 can have any other shape or color, including, but not limited to, circle, square, rectangle, star, translucent color, yellow, red or other graphical indications. As the wearer moves about the court, the trackingimages540,542 also move, following the wearer. This provides an improvement by assisting video viewers with identifying the spotlighted athlete amongst a group of other athletes.
In an embodiment illustrated inFIG. 67, thevideo generator28 is configured to generate ananimation set544 having a plurality of different animations of the trackingimages540,542. The animations vary with the athlete's actual performance, which is recorded based on the stats collected by the programmeddevice120. In the example shown, animation A (foot highlight) corresponds to a default mode, animation B1 (foot smoke) corresponds to a streak of two shots made by the tracked athlete, animation B2 (foot fire) corresponds to a streak of three shots made by the tracked athlete, animation B3 (foot blaze) corresponds to the tracked player achieving twenty points, animation C1 (foot snowflakes) corresponds to a streak of three shots missed by the tracked player, animation C2 (foot ice cubes) corresponds to over three turnovers by the tracked player, and animation C3 (foot icicles) corresponds to the tracked player having a ratio of made shots to missed shots (or shooting percentages) that is below a designated threshold.
Depending upon the embodiment, thenetwork16 can include one or more of the following: a wired network, a wireless network, an LAN, an extranet, an intranet, a WAN (including, but not limited to, the Internet), a virtual private network (“VPN”), an interconnected data path across which multiple devices may communicate, a peer-to-peer network, a telephone network, portions of a telecommunications network for sending data through a variety of different communication protocols, a Bluetooth® communication network, an RF data communication network, an IR data communication network, a satellite communication network or a cellular communication network for sending and receiving data through short messaging service (“SMS”), multimedia messaging service (“MMS”), hypertext transfer protocol (“HTTP”), direct data connection, Wireless Application Protocol (“WAP”), email or any other suitable message transfer service or format.
In an embodiment, such one or more processors (e.g., processor14) can include a data processor or a central processing unit (“CPU”). Each such one or more data storage devices can include, but is not limited to, a hard drive with a spinning magnetic disk, a Solid-State Drive (“SSD”), a floppy disk, an optical disk (including, but not limited to, a CD or DVD), a Random Access Memory (“RAM”) device, a Read-Only Memory (“ROM”) device (including, but not limited to, programmable read-only memory (“PROM”), electrically erasable programmable read-only memory (“EPROM”), electrically erasable programmable read-only memory (“EEPROM”), a magnetic card, an optical card, a flash memory device (including, but not limited to, a USB key with non-volatile memory, any type of media suitable for storing electronic instructions or any other suitable type of computer-readable storage medium. In an embodiment, an assembly includes a combination of: (a) one or more of thedatabases12 that store thesystem13; and (b) one or more of the foregoing processors (e.g., processor14).
Referring toFIG. 1, the users of thesystem13 can use or operate any suitable input/output (I/O) device to transmit inputs toprocessor14 and to receive outputs fromprocessor14, including, but not limited to, any of the devices20 (FIG. 1). Depending upon the embodiment, thedevices20 can include a personal computer (PC) (including, but not limited to, a desktop PC, a laptop or a tablet), smart television, Internet-enabled TV, person digital assistant, smartphone, cellular phone or mobile electronic device. In one embodiment, such I/O device has at least one input device (including, but not limited to, a touchscreen, a keyboard, a microphone, a sound sensor or a speech recognition device) and at least one output device (including, but not limited to, a speaker, a display screen, a monitor or an LCD).
In an embodiment, thesystem13 includes computer-readable instructions, algorithms and logic that are implemented with any suitable programming or scripting language, including, but not limited to, C, C++, Java, COBOL, assembler, PERL, Visual Basic, SQL Stored Procedures or Extensible Markup Language (XML). Thesystem13 can be implemented with any suitable combination of data structures, objects, processes, routines or other programming elements.
In an embodiment, the interfaces displayable by thedevices20 can include GUIs structured based on any suitable programming language. Each GUI can include, in an embodiment, multiple windows, pull-down menus, buttons, scroll bars, iconic images, wizards, the mouse symbol or pointer, and other suitable graphical elements. In an embodiment, the GUIs incorporate multimedia, including, but not limited to, sound, voice, motion video and virtual reality interfaces to generate outputs of thesystem13 or thedevice20.
In an embodiment, the memory devices and data storage devices described above can be non-transitory mediums that store or participate in providing instructions to a processor for execution. Such non-transitory mediums can take different forms, including, but not limited to, non-volatile media and volatile media. Non-volatile media can include, for example, optical or magnetic disks, flash drives, and any of the storage devices in any computer. Volatile media can include dynamic memory, such as main memory of a computer. Forms of non-transitory computer-readable media therefore include, for example, a floppy disk, flexible disk, hard disk, magnetic tape, any other magnetic medium, a CD-ROM, DVD, any other optical medium, punch cards, paper tape, any other physical medium with patterns of holes, a RAM, a PROM, and EPROM, a FLASH-EPROM, any other memory chip or cartridge, or any other medium from which a computer can read programming code and/or data. Many of these forms of computer readable media may be involved in carrying one or more sequences of one or more instructions to a processor for execution. In contrast with non-transitory mediums, transitory physical transmission media can include coaxial cables, copper wire and fiber optics, including the wires that comprise a bus within a computer system, a carrier wave transporting data or instructions, and cables or links transporting such a carrier wave. Carrier-wave transmission media can take the form of electric or electromagnetic signals, or acoustic or light waves such as those generated during RF and IR data communications.
It should be appreciated that at least some of the subject matter disclosed herein includes or involves a plurality of steps or procedures. In an embodiment, as described, some of the steps or procedures occur automatically or autonomously as controlled by a processor or electrical controller without relying upon a human control input, and some of the steps or procedures can occur manually under the control of a human. In another embodiment, all of the steps or procedures occur automatically or autonomously as controlled by a processor or electrical controller without relying upon a human control input. In yet another embodiment, some of the steps or procedures occur semi-automatically as partially controlled by a processor or electrical controller and as partially controlled by a human.
It should also be appreciated that aspects of the disclosed subject matter may be embodied as a method, device, assembly, computer program product or system. Accordingly, aspects of the disclosed subject matter may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.), or an embodiment combining software and hardware aspects that may all, depending upon the embodiment, generally be referred to herein as a “service,” “circuit,” “circuitry,” “module,” “assembly” and/or “system.” Furthermore, aspects of the disclosed subject matter may take the form of a computer program product embodied in one or more computer readable mediums having computer readable program code embodied thereon.
Aspects of the disclosed subject matter are described herein in terms of steps and functions with reference to flowchart illustrations and block diagrams of methods, apparatuses, systems and computer program products. It should be understood that each such step, function block of the flowchart illustrations and block diagrams, and combinations thereof, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create results and output for implementing the functions described herein.
These computer program instructions may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the functions described herein.
The computer program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions described herein.
Additional embodiments include any one of the embodiments described above, where one or more of its components, functionalities or structures is interchanged with, replaced by or augmented by one or more of the components, functionalities or structures of a different embodiment described above.
It should be understood that various changes and modifications to the embodiments described herein will be apparent to those skilled in the art. Such changes and modifications can be made without departing from the spirit and scope of the present disclosure and without diminishing its intended advantages. It is therefore intended that such changes and modifications be covered by the appended claims.
Although several embodiments of the disclosure have been disclosed in the foregoing specification, it is understood by those skilled in the art that many modifications and other embodiments of the disclosure will come to mind to which the disclosure pertains, having the benefit of the teaching presented in the foregoing description and associated drawings. It is thus understood that the disclosure is not limited to the specific embodiments disclosed herein above, and that many modifications and other embodiments are intended to be included within the scope of the appended claims. Moreover, although specific terms are employed herein, as well as in the claims which follow, they are used only in a generic and descriptive sense, and not for the purposes of limiting the present disclosure, nor the claims which follow.