Movatterモバイル変換


[0]ホーム

URL:


US10638174B2 - Synchronized video control system for sexual stimulation devices - Google Patents

Synchronized video control system for sexual stimulation devices
Download PDF

Info

Publication number
US10638174B2
US10638174B2US16/214,030US201816214030AUS10638174B2US 10638174 B2US10638174 B2US 10638174B2US 201816214030 AUS201816214030 AUS 201816214030AUS 10638174 B2US10638174 B2US 10638174B2
Authority
US
United States
Prior art keywords
video
users
video content
user
processor
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
US16/214,030
Other versions
US20200099970A1 (en
Inventor
Brian Sloan
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US16/139,550external-prioritypatent/US10576013B1/en
Application filed by IndividualfiledCriticalIndividual
Priority to US16/214,030priorityCriticalpatent/US10638174B2/en
Publication of US20200099970A1publicationCriticalpatent/US20200099970A1/en
Application grantedgrantedCritical
Priority to US16/861,014prioritypatent/US11185465B2/en
Publication of US10638174B2publicationCriticalpatent/US10638174B2/en
Priority to US16/942,751prioritypatent/US11197074B2/en
Priority to US17/534,155prioritypatent/US11478398B2/en
Priority to US17/737,968prioritypatent/US11590052B2/en
Priority to US17/737,974prioritypatent/US11607366B2/en
Priority to US17/839,426prioritypatent/US11596573B2/en
Priority to US17/853,316prioritypatent/US11540971B2/en
Priority to US17/857,983prioritypatent/US11771618B2/en
Priority to US17/857,423prioritypatent/US20220331196A1/en
Priority to US17/970,299prioritypatent/US11717467B2/en
Priority to US18/092,438prioritypatent/US20230147864A1/en
Priority to US18/178,549prioritypatent/US11779512B2/en
Priority to US18/183,948prioritypatent/US11766380B2/en
Priority to US18/185,284prioritypatent/US11759387B2/en
Priority to US18/324,983prioritypatent/US12011407B2/en
Priority to US18/365,158prioritypatent/US12213930B2/en
Priority to US18/461,413prioritypatent/US20240009072A1/en
Priority to US18/492,629prioritypatent/US20240050306A1/en
Priority to US18/666,683prioritypatent/US20240315917A1/en
Priority to US18/914,015prioritypatent/US20250032350A1/en
Priority to US18/913,973prioritypatent/US20250032349A1/en
Activelegal-statusCriticalCurrent
Anticipated expirationlegal-statusCritical

Links

Images

Classifications

Definitions

Landscapes

Abstract

A system and method for creating and sharing synchronized video control data for sexual stimulation devices that allows for automated synchronization with any video of sexual activity and allows for comprehensive personalization and customization of the video selection, synchronization, and sharing experience based on the user's preferences and biometric data.

Description

CROSS-REFERENCE TO RELATED APPLICATIONS
This application is a continuation-in-part of U.S. patent application Ser. No. 16/139,550 titled “SYNCHRONIZED VIDEO CONTROL SYSTEM FOR SEXUAL STIMULATION DEVICES”, filed on Sep. 24, 2018, the entire specifications of which is incorporated herein by reference.
BACKGROUNDField of the Art
The present invention is in the field of computer control systems, and more specifically the field of synchronized video control systems for sexual stimulation devices.
Discussion of the State of the Art
In the field of sexual stimulation devices, there are examples of control systems that allow for synchronization of the device with videos of sexual activity. However, existing systems are extremely limited in their functionality. They contain only limited libraries of manually pre-programmed synchronized stimulation routines, cannot recognize video content on their own, cannot automatically create their own stimulation routines, and cannot customize the experience for the user using biometric data about the user.
What is needed is a system for creating and sharing synchronized video control data for sexual stimulation devices that allows for automated synchronization with any video of sexual activity and allows for comprehensive personalization and customization of the video selection, synchronization, and sharing experience based on the user's preferences and biometric data.
SUMMARY
Accordingly, the inventor has conceived, and reduced to practice, a system and method for creating and sharing synchronized video control data for sexual stimulation devices that allows for automated synchronization with any video of sexual activity and allows for comprehensive personalization and customization of the video selection, synchronization, and sharing experience based on the user's preferences and biometric data.
In a preferred embodiment, a system for creating and sharing synchronized video control data for sexual stimulation devices is disclosed, comprising: a user portal comprising at least a plurality of programming instructions stored in the memory of, and operating on at least one processor of, a computing device, wherein the plurality of programming instructions, when operating on the at least one processor, cause the computing device to: allow a plurality of users to create user accounts; allow the users to choose video content from their user accounts; allow users to tag the video content with device control data; and allow users to share the tagged video content and its associated device control data to a community portal; a community portal comprising at least a plurality of programming instructions stored in the memory of, and operating on at least one processor of, a computing device, wherein the plurality of programming instructions, when operating on the at least one processor, cause the computing device to: allow a plurality of users to access and share videos and associated device control data; a video tagging system comprising at least a plurality of programming instructions stored in the memory of, and operating on at least one processor of, a computing device, wherein the plurality of programming instructions, when operating on the at least one processor, cause the computing device to: access video content containing sexual activity; allow a user to tag the video content with at least one component of activity corresponding to the sexual activity shown in the video, or choose to have the video tagged automatically by the system; generate device control data corresponding to the tagged video content; and store the tagged video content and associated device control data or transmit the tagged video content and associated device control data to a device controller; and a device controller comprising at least a plurality of programming instructions stored in the memory of, and operating on at least one processor of, a computing device, wherein the plurality of programming instructions, when operating on the at least one processor, cause the computing device to: transmit the device control data, synchronized with the video, to a compatible stimulation device such that compatible device emulates, in at least one respect, the activity shown in the video.
According to another preferred embodiment of the invention, a method for creating and sharing synchronized video control data for sexual stimulation devices is disclosed, comprising the steps of: allowing a plurality of users to create user accounts; allowing the users to choose video content from their user accounts; allowing users to tag the video content with device control data; allowing users to share the tagged video content and its associated device control data to a community portal; allowing a plurality of users to access and share videos and associated device control data; accessing video content containing sexual activity; allowing a user to tag the video content with at least one component of activity corresponding to the sexual activity shown in the video, or choose to have the video tagged automatically; generating device control data corresponding to the tagged video content; storing the tagged video content and associated device control data or transmit the tagged video content and associated device control data to a device controller; and transmitting the device control data, synchronized with the video, to a compatible stimulation device such that compatible device emulates, in at least one respect, the activity shown in the video.
BRIEF DESCRIPTION OF THE DRAWING FIGURES
The accompanying drawings illustrate several aspects and, together with the description, serve to explain the principles of the invention according to the aspects. It will be appreciated by one skilled in the art that the particular arrangements illustrated in the drawings are merely exemplary, and are not to be considered as limiting of the scope of the invention or the claims herein in any way.
FIG. 1 shows an overview of a sexual interaction and sharing platform, according to a preferred embodiment.
FIG. 2 shows additional components of a user portal, according to one aspect.
FIG. 3 shows additional components of a video selection engine, according to one aspect.
FIG. 4 shows additional components of a community portal, according to one aspect.
FIG. 5 shows an exemplary process flow for shared experience, according to one aspect.
FIG. 6 is an illustration of an exemplary shared experience user interface, according to one aspect.
FIG. 7 shows additional components of a video sharing portal, according to one aspect.
FIG. 8 shows additional components of a manual tagging system, according to one aspect.
FIG. 9 shows additional components of a manual tagging application, according to one aspect.
FIG. 10 shows additional components of an automated tagging engine, according to one aspect.
FIG. 11 shows additional components of a crowdsourcing system, according to one aspect.
FIG. 12 is a block diagram of an exemplary synchronized video control system for sexual stimulation devices according to an embodiment.
FIG. 13 is a block diagram of the video analysis engine aspect of an exemplary synchronized video control system for sexual stimulation devices according to an embodiment.
FIG. 14 is a block diagram of the control interface aspect of an exemplary synchronized video control system for sexual stimulation devices according to an embodiment.
FIG. 15 is a block diagram of the device controller aspect of an exemplary synchronized video control system for sexual stimulation devices according to an embodiment.
FIG. 16 is a flow diagram showing a method for an exemplary synchronized video control system for sexual stimulation devices according to an embodiment.
FIG. 17 is a block diagram illustrating an exemplary hardware architecture of a computing device.
FIG. 18 is a block diagram illustrating an exemplary logical architecture for a client device.
FIG. 19 is a block diagram showing an exemplary architectural arrangement of clients, servers, and external services.
FIG. 20 is another block diagram illustrating an exemplary hardware architecture of a computing device.
FIG. 21 is a flow diagram illustrating an exemplary method for a synchronized video control system for sexual stimulation devices.
DETAILED DESCRIPTION
The inventor has conceived, and reduced to practice, a system and method for creating and sharing synchronized video control data for sexual stimulation devices that allows for automated synchronization with any video of sexual activity and allows for comprehensive personalization and customization of the video selection, synchronization, and sharing experience based on the user's preferences and biometric data.
In the field of sexual stimulation devices, there are examples of control systems for such devices that allow for synchronization of the device with videos of sexual activity. These control systems attempt to replicate the sexual activities shown on the screen through synchronization of the video with some form of sexual stimulation device. However, existing systems cannot synchronize with any video of sexual activity. They must be manually pre-programmed for each video. As a result, they contain only limited libraries of video-synchronized stimulation routines. Further, since they are manually pre-programmed, the experience is the same for every user, and cannot be customized to the user's preferences or biometric data. the library of tagged videos is very small and therefore users may not find videos that contain sexual activities matching their sexual preferences for watching such videos. Further such systems cannot customize the experience for the user using biometric data about the user such as differences in anatomy. There are numerous improvements of this invention over the prior art, such as automated real time video analysis and synchronization and broad customization of the user experience based on user preferences and the user's biometric data.
In an aspect of an embodiment, the system allows users to manually tag videos containing sexual content by moving an on-screen slider bar in a manner corresponding to the movements the user sees in the video. The tagging information is stored, and may be transmitted directly to a compatible sexual stimulation device so that the user can feel in real time the actions indicated by the slider bar, and thus confirm whether the data accurately captures the movements the user sees in the video. The data from the on-screen slider can also be stored by the system for later use, or for sharing with others who have similarly compatible sexual stimulation devices.
In another aspect of an embodiment, the manual tagging data from a plurality of users may be rated by other users for accuracy and/or realism, allowing a community of users to collectively determine via a rating system which manual tagging data are most accurate and realistic for a particular video or videos.
In another aspect of an embodiment, the control system uses automated, real-time video analysis and machine learning algorithms to identify components of the sexual activity in the video such as movement, pressure, and rhythm, as opposed to existing systems which require manually pre-programming the controller to match the perceived activity in the videos. Using real-time video analysis allows access to the entirety of sexual video content available on the internet. Any video containing sexual content could be used with the system, which allows the user to choose videos with very specific content based on the user's preferences.
In another aspect of an embodiment, the manual tagging data from a plurality of users may be processed through machine learning algorithms to train the automated, real-time video analysis system to more accurately mimic the sexual content of videos. A variety of methods may be used for combining or selecting the manual tagging data to be used for such training, including the rating system described above for community rating of manual tagging data as accurate and realistic.
Since any video containing sexual content can be used, the system can be tuned to mimic the sexual activity of particular actors or actresses engaging in specific sexual activities. The machine learning algorithms used to conduct the video analysis could be fed metadata about the videos such as the names of the actor or actress, or perform facial recognition to identify videos containing certain actors and actresses, such that the control system could learn to “perform” certain sexual activities just like a certain actor or actress does generally, or even in a particular film. Likewise, since any video can be used with this control system, and since the synchronization with the videos can be either automated or controlled by the user, the opportunities for customization and sharing are unlimited. Users would have the opportunity to customize the stimulation associated with videos in myriad ways, and share those customized experiences with others by sharing the control system file associated with that video. Further, users could create videos of their own sexual activity, and “share” their experience with others remotely through this control system and an appropriate stimulation device.
Another major benefit of this control system is the broad customization allowed based on user profiles. At the most basic level, users can simply watch a video, and allow the system to control the device based on the system's automated video parsing without any adjustment or input. However, the system is not limited to such usage. The system could be used with a manual form of input such as a slider bar on the screen, which allows users to map their own perception of the movement and sexual activity in the video.
The system could allow users to set up a profile containing parameters and preferences for operation of the compatible sexual stimulation device. For example, the user could set a parameter indicating that the device should speed up or slow down when certain movements in the video are detected. Further, combining these parameters with biometric sensor data could allow the user to indicate that the device should attempt to prolong orgasm for a certain period of time. For example, the user could set a parameter indicating that orgasm should be delayed at least 15 minutes, and this parameter, combined with biometric data of breathing rate, heart rate, penile stiffness, etc., could cause the control system to slow down or stop stimulation until the biometric data falls back within certain ranges, at which point the device would continue stimulation as usual. A myriad of parameters, preferences, and biometric data ranges could be used. For example, the control system could be instructed to delay orgasm, prevent orgasm, or hold the user at a given level of excitement.
Further, metadata can be captured from the video related to the video content including, for example, the actor or actress in the video, the type of sexual activity, the position or orientation of the sexual activity, the location or scene in which the sexual activity occurs, and the style or category of the video content (e.g., oral sex, anal sex, gay sex, fetish). Using metadata associated with the videos, the control system could select or suggest videos containing very specific content based on the user's preferences. Such metadata may already be embedded in the videos, may be available on the internet, or may be developed by having users input such metadata in a growing library of such videos. Further, biometric data, for example penis length and girth for males, and vaginal temperature and vaginal blood flow for females, can be entered into the user profile, and the stimulation provided by the control system can be automatically adjusted to provide the user a customized, better feeling, more realistic experience based on those dimensions. For females, the amount of vaginal secretions could be measured using sensors on a compatible device, and the compatible device's operation could be adjusted accordingly. Optionally, other types of biometric data such as heart rate, breathing rate, and penile stiffness could be captured by a variety of commercially available devices (for example, sports training monitors), or by sensors on the stimulation device, itself, and fed back to the user profile to automatically optimize the video content and types of stimulation preferred by the user.
In another aspect of an embodiment, the system could provide the user with a wider variety of videos that match the user's preferences. For example, users could manually tag a small subset of videos with synchronized stimulation routines, which could then be applied by the machine learning algorithms to very large databases of videos to learn which videos contain that sort of sexual activity. Clustering could be used to identify certain types of sexual activity preferred by the user, based on the movement, rhythm, and pressure associated with them. User ratings in some portal or online platform could help refine the video selection outputs and extrapolations generated by the machine learning algorithms.
In some embodiments, all components of the video control system may be located on a general purpose computer. In other embodiments, some components of the video control system may be located on the compatible stimulation device as embedded computer components or systems. For example, a compatible stimulation device may contain an embedded computer component or systems that act as the device controller, which receives signals from a video analysis engine and causes the compatible stimulation device to operate in accordance with those signals. In some aspects of some embodiments, such an embedded computer component or system might contain programmed sequences of movements or other content such that the bandwidth required to transmit signals to the device can be reduced by sending references to the programmed sequences of movements.
Definitions
The terms “tag” and “tagging” as used herein mean the annotation or encoding of video files with additional information that allows for the control of compatible stimulation devices in synchronization with activities in the video. The annotation may be added to the video file itself, or may be a separate file or files that contain information related to the video.
Conceptual Architecture
FIG. 1 shows an overview of a sexual interaction and sharing platform100, according to a preferred embodiment. According to the embodiment, a sexual interaction and sharing platform100 may comprise a user portal150 the stores and provides user profiles, preferences, and other data as well as maintains lists of trusted or shared users and other associations. Acontent manager110 may be used to handle the storage and retrieval of media content, such as images and videos. Acommunity portal120 may be used to provide a trust or rating system that may be used to manage user relationships and associations and provide the groundwork for a variety of shared user spaces such as a video sharing portal for watching content with other users, a private session portal for sharing experiences with specific users in a private setting, themed areas that may provide experiences catering to specific topics or styles, and discussion boards or other social networking features. Acrowdsourcing engine140 may be used to provide additional content to users such as newly-created or exclusive content, rewards programs such as actual money or proprietary tokens or other benefits, as well as online marketplaces and vendors that may provide a variety of paid content or products, and from which the platform100 may receive a fee or percentage. A video analysis and control engine130 may be used to provide an application portal for manually tagging media as well as an artificial intelligence (AI) driven automated tagging engine, as well as storing videos and their associated tags and other metadata and providing features for remote control of stimulation devices for shared experiences.
FIG. 2 shows additional components of a user portal150, according to one aspect. According to the aspect, a user portal150 may comprise acentral storage155 for user information, including (but not limited to) user profiles152, preferences153 such as preferences for application or device settings or themed areas, or user biometric data154 such as various metrics that may be collected by devices for or during use, such as (for example) penis girth or length, body temperature, heart rate, and other biometric data. Anaccount anonymizer151 may be utilized to anonymize accounts for privacy reasons, such as by obfuscating personally-identifiable details such as name or location. A trusted friend and sharinglist156 may be used to maintain user associations and relationships, such as users that have added each other as “friends” in a discussion board, or have participated in group media viewing or shared experiences. User data fromstorage155 may be provided for use by other components for operation, such as avideo selection engine160 that may select and retrieve videos according to a user's preferences and provide them to a video analysis engine, a device adjustment driver157 that may direct remote-operated stimulation devices and provide control commands to a control interface, and acommunity portal interface158 that may incorporate user details into a community setting for socialization and other interactions. Data encryption may be used to protect sensitive user data such asuser profile152, user preferences153, and user biometrics154.
FIG. 3 shows additional components of avideo selection engine160, according to one aspect. According to the aspect, asearch engine161 may receive user preferences from auser portal160, and utilize these preferences to direct the operation of aweb crawler162 to locate and retrieve media content over a network (such as content from online streaming service providers, or publicly-posted content on social networking sites).Search engine161 may incorporate information from avideo metadata library165 to appropriately locate content in accordance with the user preferences received, searching for and retrieving contents based on whether it matches metadata specified in the preferences. Avideo retriever163 may be used to retrieve stored content from a video library166 (such as a local repository of content, for example paid content exclusive to platform subscribers), as well as loading information from adevice control library164 to properly direct any stimulation devices, before providing the retrieved media content (whether retrieved fromvideo library166 or online sources via web crawler162) and device control information as video and control output.
FIG. 4 shows additional components of acommunity portal120, according to one aspect. According to the aspect, acommunity portal158 may be used to provide a trust orrating system406 that may be used to manage user relationships and associations and provide the groundwork for a variety of shared user spaces such as avideo sharing portal401 for watching content with other users, an open group sharing area402, aprivate session portal405 for sharing experiences with specific users in a private setting,themed areas404 that may provide experiences catering to specific topics or styles, and discussion boards or other social networking features403.
FIG. 7 shows additional components of avideo sharing portal401, according to one aspect. According to the aspect, user-createdvideos701 as well as non-user-created videos702 (for example, paid content or streaming content from an online service provider or public content provider such as an online discussion board) may be retrieved and processed for metadata tagging703. Tagging may occur manually, wherein a user selects metadata for association with a video, or automatically wherein AI is used to process the video and apply metadata tags based on analysis of the video details and content. The video may then be tagged withmotion data704, again either manually or automatically, comprising information on motions and activities taking place in the video that may be used to direct stimulation devices for replicating the actions happening in the video. Processed videos with metadata and motion tagging information associated with them may then be provided forcommunity availability705, enabling users to select them for group viewing and shared experiences.
FIG. 8 shows additional components of amanual tagging system800, according to one aspect. According to the aspect, amanual tagging application802 may be used to enableuser input801 to enable the user to manually associate metadata tags with videos, such as keywords and indicators that may be used to assist AI and other users in video selection and retrieval. Videos may be stored166 separately from ametadata store165, with associative links ensuring that when a video is retrieved the corresponding metadata is also retrieved while providing the benefits of distributed storage such as redundancy and decreased latency. Video-specific control data may also be stored803, comprising control information for use in directing stimulation devices during viewing of a video and comprising information from motion tagging704.
FIG. 9 shows additional components of amanual tagging application900, according to one aspect. According to the aspect, amanual tagging application900 may provide an interface for viewing avideo901 along with a plurality of controls such asinteractive sliders902 for manually indicating motion and activity in the video throughout playback (that is, motion inputs may be provided at any given point during playback, so that a continuous and changing stream of motion data is associated with the video as it plays). Adata collector display903 may collect and present metadata statistics such as motion intensity or number of tagged keywords at various points in a video, giving a user a clear indication of areas where further tagging may be needed as well as assisting the user with ensuring that motion and other metadata remains consistent relative to tagged data in the rest of the video (for example, so that the user does not accidentally cause sudden and jarring changes in the motion controls, instead referencing each data input on prior inputs).
FIG. 10 shows additional components of anautomated tagging engine1000, according to one aspect. According to the aspect, AI-driven tagging may be used to augment or replace manual tagging, for example to provide a metadata baseline on all processed videos to which users may then choose to add manual tagging data. Avideo player1001 may retrieve and load a video for processing, and afacial recognition engine1002 may then analyze the video frames to identify faces and tag specific actors or other keywords based on the analysis, as well as to enforce privacy policies by (for example) obfuscating faces (such as by applying a blur to the face within each frame of a video) of some or all actors in a video.Facial component parser1003 may be used for more fine-grained detail in tagging, for example discriminating between specific emotions or by identifying facial features such as eyes, lips, nose, etc. for use in motion tracking. Amotion tracking engine1004 may then analyze movement in the video to identify actions and motions that are occurring, loading reference data from amotion library1005 to improve results and maintain consistency across videos and using avideo timing module1006 to ensure the tagged motion data closely matches the actions occurring in the video. Taggedvideo data1007 may then be stored and provided as output for further use.
FIG. 11 shows additional components of acrowdsourcing system1100, according to one aspect. According to the aspect, crowdsourcing may comprise a closed-loop system for users to both produce and consume content, with a two-way flow of rewards and payments to provide a full virtualized marketplace. User profiles152 may provide content to anonline storefront1110 for purchase (for example, to sell user-generated content or to sell services such as metadata or motion tagging for other videos), as well as to acommunity portal120 for non-paid content. This enables users to build a portfolio of both free and paid content, while incentivizing both. Content in acommunity portal120 may generate rewards for theuser152, either incentivizing an initial upload by giving reward points or tokens on a per-content basis, or providing rewards based on community engagement to incentivize content on a per-view basis. Paid content in astorefront1110 may generate actual revenue in the form of received payments, optionally with the crowdsourcing platform receiving a portion to monetize user content. Users interacting via a community portal may be offered content fromstorefront1110, and may purchase content from thecommunity portal120 directly (that is, without needing to be redirected to an external virtual marketplace).
Detailed Description of Exemplary Aspects
FIG. 5 shows an exemplary process flow for sharedexperience500, according to one aspect. According to the aspect, a user first uploads a tagged video501 tostorage502, or otherwise selects or links to a tagged video501 fromstorage502, comprising both media content and metadata (the associated tags). Fromstorage502, this content may then be accessed503 by group participants via acommunity portal120 as described above. The video content may then be synchronized to each user'sdevice510, for simultaneous stimulation by each user on theirrespective devices511,512,513 in synchronization with each other user. Users may also be provided a shared video feed520, showing a video feed of each other participating user in a manner similar to videoconferencing calls, providing the shared experience of watching the video together and allowing for visual or auditory interaction between users during the shared experience.
FIG. 6 is an illustration of an exemplary sharedexperience user interface600, according to one aspect. According to the aspect, a user interface for a sharedexperience600 may show a given user a display of the media content selected for viewing601, as well as video feeds of other participatingusers602,603, providing for user interaction as well as the simulation of a shared, group viewing experience. A user may also be shown a manualdevice control override604, enabling them to take manual control of any stimulation devices in use rather than allowing AI-driven operation to direct them during the experience (as described below).
FIG. 12 is a block diagram of an exemplary synchronized video control system forsexual stimulation devices1200 according to a preferred embodiment. In this embodiment, avideo analysis engine1201 inputs a video of sexual activity, parses the video into at least the components of movement corresponding to the sexual activity shown in the video, and outputs signals containing the parsed video information to adevice controller1202. Auser interface1203 allows the user to enter a profile containing at least his preferences regarding the video content that he prefers, stores the user's profile information, and outputs the user's profile information to thedevice controller1202. Thedevice controller1202 adjusts the signals from thevideo analysis engine1201 based on the profile information from theuser interface1203 and outputs the adjusted signals to astimulation device1204 such that they are synchronized with the activity shown in the video. In an aspect of an embodiment, the parsed video information from thevideo analysis engine1201 is stored in adata storage device1205 for later retrieval and use.
FIG. 13 is a block diagram1300 thevideo analysis engine1201 aspect of an exemplary synchronized video control system for sexual stimulation devices according to a preferred embodiment. Avideo parser1301 receives video input1302, sends the video's metadata to ametadata processor1303, which checks to see if the metadata for that video already exists in thedata storage device1205. If the metadata already exists, it is read from thedata storage device1205 and sent out theuser interface1203. If the metadata does not exist, it is formatted, written to thedata storage device1205, and sent out to theuser interface1203. Simultaneously, the video parser1302 sends the video content to themotion translation processor1304, which checks to see if the control signal data for that video already exists in thedata storage device1205, and incorporates data from thecontrol interface1305. If the control signal data already exists, it is read from thedata storage device1205 and sent out thedevice controller1202. If the control signals do not exist, themotion translation processor1304 uses video processing algorithms and machine learning algorithms to detect sexual activity and to translate the motions in the video to control signals related to movement, pressure, and rhythm. The controls signals are then written to thedata storage device1205 and sent out to thedevice controller1202. In an aspect of an embodiment, the actual video content may also be stored in thedata storage device1205.
FIG. 14 is a block diagram1400 of theuser interface1203 aspect of an exemplary synchronized video control system for sexual stimulation devices according to a preferred embodiment. Through thevideo preference interface1401, the user can either select videos manually, or enter preferences regarding the types of video content he prefers based on metadata associated with the videos. The user can further enter biometric data manually, or it may be obtained automatically by thebiometric data interface1402 frombiometric sensor receiver1504 disclosed inFIG. 15. The biometric data is sent to a penilesize conversion calculator1403, which makes adjustments to output signals to provide a similar experience to the user regardless of penis size. The video preferences, biometric data, and penile size conversions are sent to aprofile generator1404, which creates a profile for the user based on the various inputs. The profile information is saved to thestorage device1205, and is sent to thedevice controller1202. The user interface may contain a manual video tagging interface1405, which allows the user to adjust the sensations received while viewing those videos. The user interface may also contain a video sharing portal1406 that allows for the storage and sharing of a library of analyzed videos such that users can further customize their experiences by, for example, selecting certain experiences based on their video preferences, sharing personalized experiences by sharing their own videos, sharing user profiles, or sharing customized parsing of certain videos.
FIG. 15 is a block diagram1500 of thedevice controller1202 aspect of an exemplary synchronized video control system for sexual stimulation devices according to a preferred embodiment. Control signals for the video being watched are received from themotion translation processor1304 into thevideo synchronizer1501, which adjusts the timing of the signals to correspond with the video being watched. Biometric and preference data is received into theprofile interface1502 from theprofile generator1404. Acontrol signal generator1503 receives the outputs from both thevideo synchronizer1501 andprofile interface1502, and adjusts the synchronized control signals based on the biometric and preference data, and sends out the adjusted control signal to thestimulation device1204. The device controller may also contain abiometric sensor receiver1504 that could allow the capture of biometric data from wireless devices such as fitness trackers that monitor heart rate, blood pressure and breathing monitors, and even a penile stiffness sensor in the stimulation device itself. The data captured through the biometric sensor receiver could be used for real time feedback to thecontrol signal generator1503 and for use in improving user experiences by enhancing the user's profile or improving the accuracy of video selection.
FIG. 16 is a flow diagram showing amethod1600 for an exemplary synchronized video control system for sexual stimulation devices according to a preferred embodiment. According to this method, video of sexual activity would be input into acomputer1601. The computer, using machine learning algorithms, would parse the video into at least the components of movement corresponding to the sexual activity shown in thevideo1602. The parsed video information is stored1603, and the video metadata is also stored1604. The parsed video information is simultaneously converted to device control signals and output to a device controller1605. Separately, information about theuser device parameters1606 may be entered manually or captured by querying the device itself.Biometric data1607 may also be entered manually or captured from wireless devices such as fitness trackers that monitor heart rate, blood pressure and breathing monitors, and even a penile stiffness sensor in the stimulation device itself. Thedevice parameters1606 andbiometric data1607 are stored as part of the user'sprofile information1608, and then output to adevice controller1609. The output signals from the parsed video information1605 are adjusted1610 based on the output from theprofile information1609, and output to a stimulation device such that they are synchronized with the activity shown in the video such that a stimulation device capable of receiving the signals from the device controller and translating the signals into sexual stimulation would emulate the sexual activity shown in thevideo1611. In some embodiments, the user may also share with other users his or her user profile, personalized experiences by sharing his own videos, or customized parsing of certain videos.
FIG. 21 is a flow diagram illustrating an exemplary method for a synchronized video control system for sexual stimulation devices, comprising the steps of: allowing a plurality of users to createuser accounts2101, allowing the users to choose or access video content from theiruser accounts2102, allowing users to tag the video content with data corresponding to at least one component of activity corresponding to the sexual activity shown in the video, or choose to have the video tagged automatically2103, generating device control data corresponding to the taggedvideo content2104, Store the tagged video content and associated device control data or transmit the tagged video content and associated device control data to adevice controller2105, allowing a plurality of users to access and share videos and associated device control data2106, and transmitting the device control data, synchronized with a video, to a compatible stimulation device such that compatible device emulates, in at least one respect, the activity shown in thevideo2107.
Hardware Architecture
Generally, the techniques disclosed herein may be implemented on hardware or a combination of software and hardware. For example, they may be implemented in an operating system kernel, in a separate user process, in a library package bound into network applications, on a specially constructed machine, on an application-specific integrated circuit (ASIC), or on a network interface card.
Software/hardware hybrid implementations of at least some of the aspects disclosed herein may be implemented on a programmable network-resident machine (which should be understood to include intermittently connected network-aware machines) selectively activated or reconfigured by a computer program stored in memory. Such network devices may have multiple network interfaces that may be configured or designed to utilize different types of network communication protocols. A general architecture for some of these machines may be described herein in order to illustrate one or more exemplary means by which a given unit of functionality may be implemented. According to specific aspects, at least some of the features or functionalities of the various aspects disclosed herein may be implemented on one or more general-purpose computers associated with one or more networks, such as for example an end-user computing device, a client computer, a network server or other server system, a mobile computing device (e.g., tablet computing device, mobile phone, smartphone, laptop, or other appropriate computing device), a consumer electronic device, a music player, or any other suitable electronic device, router, switch, or other suitable device, or any combination thereof. In at least some aspects, at least some of the features or functionalities of the various aspects disclosed herein may be implemented in one or more virtualized computing environments (e.g., network computing clouds, virtual machines hosted on one or more physical computing machines, or other appropriate virtual environments).
Referring now toFIG. 17, there is shown a block diagram depicting anexemplary computing device10 suitable for implementing at least a portion of the features or functionalities disclosed herein.Computing device10 may be, for example, any one of the computing machines listed in the previous paragraph, or indeed any other electronic device capable of executing software- or hardware-based instructions according to one or more programs stored in memory.Computing device10 may be configured to communicate with a plurality of other computing devices, such as clients or servers, over communications networks such as a wide area network a metropolitan area network, a local area network, a wireless network, the Internet, or any other network, using known protocols for such communication, whether wireless or wired.
In one aspect,computing device10 includes one or more central processing units (CPU)12, one ormore interfaces15, and one or more busses14 (such as a peripheral component interconnect (PCI) bus). When acting under the control of appropriate software or firmware, CPU12 may be responsible for implementing specific functions associated with the functions of a specifically configured computing device or machine. For example, in at least one aspect, acomputing device10 may be configured or designed to function as a server system utilizing CPU12,local memory11 and/orremote memory16, and interface(s)15. In at least one aspect, CPU12 may be caused to perform one or more of the different types of functions and/or operations under the control of software modules or components, which for example, may include an operating system and any appropriate applications software, drivers, and the like.
CPU12 may include one ormore processors13 such as, for example, a processor from one of the Intel, ARM, Qualcomm, and AMD families of microprocessors. In some aspects,processors13 may include specially designed hardware such as application-specific integrated circuits (ASICs), electrically erasable programmable read-only memories (EEPROMs), field-programmable gate arrays (FPGAs), and so forth, for controlling operations ofcomputing device10. In a particular aspect, a local memory11 (such as non-volatile random access memory (RAM) and/or read-only memory (ROM), including for example one or more levels of cached memory) may also form part of CPU12. However, there are many different ways in which memory may be coupled tosystem10.Memory11 may be used for a variety of purposes such as, for example, caching and/or storing data, programming instructions, and the like. It should be further appreciated that CPU12 may be one of a variety of system-on-a-chip (SOC) type hardware that may include additional hardware such as memory or graphics processing chips, such as a QUALCOMM SNAPDRAGON™ or SAMSUNG EXYNOS™ CPU as are becoming increasingly common in the art, such as for use in mobile devices or integrated devices.
As used herein, the term “processor” is not limited merely to those integrated circuits referred to in the art as a processor, a mobile processor, or a microprocessor, but broadly refers to a microcontroller, a microcomputer, a programmable logic controller, an application-specific integrated circuit, and any other programmable circuit.
In one aspect, interfaces15 are provided as network interface cards (NICs). Generally, NICs control the sending and receiving of data packets over a computer network; other types ofinterfaces15 may for example support other peripherals used withcomputing device10. Among the interfaces that may be provided are Ethernet interfaces, frame relay interfaces, cable interfaces, DSL interfaces, token ring interfaces, graphics interfaces, and the like. In addition, various types of interfaces may be provided such as, for example, universal serial bus (USB), Serial, Ethernet, FIREWIRE™, THUNDERBOLT™, PCI, parallel, radio frequency (RF), BLUETOOTH™, near-field communications (e.g., using near-field magnetics), 802.11 (WiFi), frame relay, TCP/IP, ISDN, fast Ethernet interfaces, Gigabit Ethernet interfaces, Serial ATA (SATA) or external SATA (ESATA) interfaces, high-definition multimedia interface (HDMI), digital visual interface (DVI), analog or digital audio interfaces, asynchronous transfer mode (ATM) interfaces, high-speed serial interface (HSSI) interfaces, Point of Sale (POS) interfaces, fiber data distributed interfaces (FDDIs), and the like. Generally,such interfaces15 may include physical ports appropriate for communication with appropriate media. In some cases, they may also include an independent processor (such as a dedicated audio or video processor, as is common in the art for high-fidelity A/V hardware interfaces) and, in some instances, volatile and/or non-volatile memory (e.g., RAM).
Although the system shown inFIG. 17 illustrates one specific architecture for acomputing device10 for implementing one or more of the aspects described herein, it is by no means the only device architecture on which at least a portion of the features and techniques described herein may be implemented. For example, architectures having one or any number ofprocessors13 may be used, andsuch processors13 may be present in a single device or distributed among any number of devices. In one aspect, asingle processor13 handles communications as well as routing computations, while in other aspects a separate dedicated communications processor may be provided. In various aspects, different types of features or functionalities may be implemented in a system according to the aspect that includes a client device (such as a tablet device or smartphone running client software) and server systems (such as a server system described in more detail below).
Regardless of network device configuration, the system of an aspect may employ one or more memories or memory modules (such as, for example,remote memory block16 and local memory11) configured to store data, program instructions for the general-purpose network operations, or other information relating to the functionality of the aspects described herein (or any combinations of the above). Program instructions may control execution of or comprise an operating system and/or one or more applications, for example.Memory16 ormemories11,16 may also be configured to store data structures, configuration data, encryption data, historical system operations information, or any other specific or generic non-program information described herein.
Because such information and program instructions may be employed to implement one or more systems or methods described herein, at least some network device aspects may include nontransitory machine-readable storage media, which, for example, may be configured or designed to store program instructions, state information, and the like for performing various operations described herein. Examples of such nontransitory machine-readable storage media include, but are not limited to, magnetic media such as hard disks, floppy disks, and magnetic tape; optical media such as CD-ROM disks; magneto-optical media such as optical disks, and hardware devices that are specially configured to store and perform program instructions, such as read-only memory devices (ROM), flash memory (as is common in mobile devices and integrated systems), solid state drives (SSD) and “hybrid SSD” storage drives that may combine physical components of solid state and hard disk drives in a single hardware device (as are becoming increasingly common in the art with regard to personal computers), memristor memory, random access memory (RAM), and the like. It should be appreciated that such storage means may be integral and non-removable (such as RAM hardware modules that may be soldered onto a motherboard or otherwise integrated into an electronic device), or they may be removable such as swappable flash memory modules (such as “thumb drives” or other removable media designed for rapidly exchanging physical storage devices), “hot-swappable” hard disk drives or solid state drives, removable optical storage discs, or other such removable media, and that such integral and removable storage media may be utilized interchangeably. Examples of program instructions include both object code, such as may be produced by a compiler, machine code, such as may be produced by an assembler or a linker, byte code, such as may be generated by for example a JAVA™ compiler and may be executed using a Java virtual machine or equivalent, or files containing higher level code that may be executed by the computer using an interpreter (for example, scripts written in Python, Perl, Ruby, Groovy, or any other scripting language).
In some aspects, systems may be implemented on a standalone computing system. Referring now toFIG. 18, there is shown a block diagram depicting a typical exemplary architecture of one or more aspects or components thereof on a standalone computing system.Computing device20 includesprocessors21 that may run software that carry out one or more functions or applications of aspects, such as for example aclient application24.Processors21 may carry out computing instructions under control of anoperating system22 such as, for example, a version of MICROSOFT WINDOWS™ operating system, APPLE macOS™ or iOS™ operating systems, some variety of the Linux operating system, ANDROID™ operating system, or the like. In many cases, one or more sharedservices23 may be operable insystem20, and may be useful for providing common services toclient applications24.Services23 may for example be WINDOWS™ services, user-space common services in a Linux environment, or any other type of common service architecture used withoperating system21.Input devices28 may be of any type suitable for receiving user input, including for example a keyboard, touchscreen, microphone (for example, for voice input), mouse, touchpad, trackball, or any combination thereof.Output devices27 may be of any type suitable for providing output to one or more users, whether remote or local tosystem20, and may include for example one or more screens for visual output, speakers, printers, or any combination thereof.Memory25 may be random-access memory having any structure and architecture known in the art, for use byprocessors21, for example to run software.Storage devices26 may be any magnetic, optical, mechanical, memristor, or electrical storage device for storage of data in digital form (such as those described above, referring toFIG. 17). Examples ofstorage devices26 include flash memory, magnetic hard drive, CD-ROM, and/or the like.
In some aspects, systems may be implemented on a distributed computing network, such as one having any number of clients and/or servers. Referring now toFIG. 19, there is shown a block diagram depicting anexemplary architecture30 for implementing at least a portion of a system according to one aspect on a distributed computing network. According to the aspect, any number ofclients33 may be provided. Eachclient33 may run software for implementing client-side portions of a system; clients may comprise asystem20 such as that illustrated inFIG. 18. In addition, any number ofservers32 may be provided for handling requests received from one ormore clients33.Clients33 andservers32 may communicate with one another via one or moreelectronic networks31, which may be in various aspects any of the Internet, a wide area network, a mobile telephony network (such as CDMA or GSM cellular networks), a wireless network (such as WiFi, WiMAX, LTE, and so forth), or a local area network (or indeed any network topology known in the art; the aspect does not prefer any one network topology over any other).Networks31 may be implemented using any known network protocols, including for example wired and/or wireless protocols.
In addition, in some aspects,servers32 may callexternal services37 when needed to obtain additional information, or to refer to additional data concerning a particular call. Communications withexternal services37 may take place, for example, via one ormore networks31. In various aspects,external services37 may comprise web-enabled services or functionality related to or installed on the hardware device itself. For example, in one aspect whereclient applications24 are implemented on a smartphone or other electronic device,client applications24 may obtain information stored in aserver system32 in the cloud or on anexternal service37 deployed on one or more of a particular enterprise's or user's premises.
In some aspects,clients33 or servers32 (or both) may make use of one or more specialized services or appliances that may be deployed locally or remotely across one ormore networks31. For example, one ormore databases34 may be used or referred to by one or more aspects. It should be understood by one having ordinary skill in the art thatdatabases34 may be arranged in a wide variety of architectures and using a wide variety of data access and manipulation means. For example, in various aspects one ormore databases34 may comprise a relational database system using a structured query language (SQL), while others may comprise an alternative data storage technology such as those referred to in the art as “NoSQL” (for example, HADOOP CASSANDRA™, GOOGLE BIGTABLE™, and so forth). In some aspects, variant database architectures such as column-oriented databases, in-memory databases, clustered databases, distributed databases, or even flat file data repositories may be used according to the aspect. It will be appreciated by one having ordinary skill in the art that any combination of known or future database technologies may be used as appropriate, unless a specific database technology or a specific arrangement of components is specified for a particular aspect described herein. Moreover, it should be appreciated that the term “database” as used herein may refer to a physical database machine, a cluster of machines acting as a single database system, or a logical database within an overall database management system. Unless a specific meaning is specified for a given use of the term “database”, it should be construed to mean any of these senses of the word, all of which are understood as a plain meaning of the term “database” by those having ordinary skill in the art.
Similarly, some aspects may make use of one ormore security systems36 andconfiguration systems35. Security and configuration management are common information technology (IT) and web functions, and some amount of each are generally associated with any IT or web systems. It should be understood by one having ordinary skill in the art that any configuration or security subsystems known in the art now or in the future may be used in conjunction with aspects without limitation, unless aspecific security36 orconfiguration system35 or approach is specifically required by the description of any specific aspect.
FIG. 20 shows an exemplary overview of acomputing device40 as may be used in any of the various locations throughout the system. It is exemplary of any computer that may execute code to process data. Various modifications and changes may be made tocomputing device40 without departing from the broader scope of the system and method disclosed herein. Central processor unit (CPU)41 is connected tobus42, to which bus is also connectedmemory43,nonvolatile memory44,display47, input/output (I/O)unit48, and network interface card (NIC)53. I/O unit48 may, typically, be connected tokeyboard49, pointingdevice50,hard disk52, and real-time clock51.NIC53 connects to network54, which may be the Internet or a local network, which local network may or may not have connections to the Internet. Also shown as part ofsystem40 ispower supply unit45 connected, in this example, to a main alternating current (AC)supply46. Not shown are batteries that could be present, and many other devices and modifications that are well known but are not applicable to the specific novel functions of the current system and method disclosed herein. It should be appreciated that some or all components illustrated may be combined, such as in various integrated applications, for example Qualcomm or Samsung system-on-a-chip (SOC) devices, or whenever it may be appropriate to combine multiple capabilities or functions into a single hardware device (for instance, in mobile devices such as smartphones, video game consoles, in-vehicle computing devices such as navigation or multimedia systems in automobiles, or other integrated hardware devices).
In various aspects, functionality for implementing systems or methods of various aspects may be distributed among any number of client and/or server components. For example, various software modules may be implemented for performing various functions in connection with the system of any particular aspect, and such modules may be variously implemented to run on server and/or client components.
The skilled person will be aware of a range of possible modifications of the various aspects described above. Accordingly, the present invention is defined by the claims and their equivalents.
One or more different aspects may be described in the present application. Further, for one or more of the aspects described herein, numerous alternative arrangements may be described; it should be appreciated that these are presented for illustrative purposes only and are not limiting of the aspects contained herein or the claims presented herein in any way. One or more of the arrangements may be widely applicable to numerous aspects, as may be readily apparent from the disclosure. In general, arrangements are described in sufficient detail to enable those skilled in the art to practice one or more of the aspects, and it should be appreciated that other arrangements may be utilized and that structural, logical, software, electrical and other changes may be made without departing from the scope of the particular aspects. Particular features of one or more of the aspects described herein may be described with reference to one or more particular aspects or figures that form a part of the present disclosure, and in which are shown, by way of illustration, specific arrangements of one or more of the aspects. It should be appreciated, however, that such features are not limited to usage in the one or more particular aspects or figures with reference to which they are described. The present disclosure is neither a literal description of all arrangements of one or more of the aspects nor a listing of features of one or more of the aspects that must be present in all arrangements.
Headings of sections provided in this patent application and the title of this patent application are for convenience only, and are not to be taken as limiting the disclosure in any way.
Devices that are in communication with each other need not be in continuous communication with each other, unless expressly specified otherwise. In addition, devices that are in communication with each other may communicate directly or indirectly through one or more communication means or intermediaries, logical or physical.
A description of an aspect with several components in communication with each other does not imply that all such components are required. To the contrary, a variety of optional components may be described to illustrate a wide variety of possible aspects and in order to more fully illustrate one or more aspects. Similarly, although process steps, method steps, algorithms or the like may be described in a sequential order, such processes, methods and algorithms may generally be configured to work in alternate orders, unless specifically stated to the contrary. In other words, any sequence or order of steps that may be described in this patent application does not, in and of itself, indicate a requirement that the steps be performed in that order. The steps of described processes may be performed in any order practical. Further, some steps may be performed simultaneously despite being described or implied as occurring non-simultaneously (e.g., because one step is described after the other step). Moreover, the illustration of a process by its depiction in a drawing does not imply that the illustrated process is exclusive of other variations and modifications thereto, does not imply that the illustrated process or any of its steps are necessary to one or more of the aspects, and does not imply that the illustrated process is preferred. Also, steps are generally described once per aspect, but this does not mean they must occur once, or that they may only occur once each time a process, method, or algorithm is carried out or executed. Some steps may be omitted in some aspects or some occurrences, or some steps may be executed more than once in a given aspect or occurrence.
When a single device or article is described herein, it will be readily apparent that more than one device or article may be used in place of a single device or article. Similarly, where more than one device or article is described herein, it will be readily apparent that a single device or article may be used in place of the more than one device or article.
The functionality or the features of a device may be alternatively embodied by one or more other devices that are not explicitly described as having such functionality or features. Thus, other aspects need not include the device itself.
Techniques and mechanisms described or referenced herein will sometimes be described in singular form for clarity. However, it should be appreciated that particular aspects may include multiple iterations of a technique or multiple instantiations of a mechanism unless noted otherwise. Process descriptions or blocks in figures should be understood as representing modules, segments, or portions of code which include one or more executable instructions for implementing specific logical functions or steps in the process. Alternate implementations are included within the scope of various aspects in which, for example, functions may be executed out of order from that shown or discussed, including substantially concurrently or in reverse order, depending on the functionality involved, as would be understood by those having ordinary skill in the art.

Claims (18)

What is claimed is:
1. A system for creating and sharing synchronized video control data for sexual stimulation devices, comprising:
one or more networked computing devices, each comprising a memory, one or more processors, and a plurality of programming instructions comprising an operating system; and
a user portal comprising at least a plurality of programming instructions stored in the memory of, and operating on at least one processor of, the one or more networked computing devices, wherein the plurality of programming instructions, when operating on the at least one processor, cause the one or more networked computing devices to:
allow a plurality of users to create user accounts;
allow the users to choose video content from their user accounts;
allow users to tag the video content using a video tagging system; and
allow users to share the tagged video content and its associated device control data to a community portal;
a community portal comprising at least a plurality of programming instructions stored in the memory of, and operating on at least one processor of, the one or more networked computing devices, wherein the plurality of programming instructions, when operating on the at least one processor, cause the one or more networked computing devices to:
allow a plurality of users to access and share videos and associated device control data;
a video tagging system comprising at least a plurality of programming instructions stored in the memory of, and operating on at least one processor of, the one or more networked computing devices, wherein the plurality of programming instructions, when operating on the at least one processor, cause the one or more networked computing devices to:
access video content containing depictions of sexual activity;
allow a user to tag the video content with at least one component corresponding to the depictions of sexual activity shown in the video by manually moving an on-screen object in a synchronized movement with one or more of the movements corresponding to the depictions of sexual activity shown in the video;
generate device control data corresponding to the tagged video content; and
store the tagged video content and associated device control; and
a device controller comprising at least a plurality of programming instructions stored in the memory of, and operating on at least one processor of, the one or more networked computing devices, wherein the plurality of programming instructions, when operating on the at least one processor, cause the one or more networked computing devices to:
transmit the device control data, synchronized with the video, to a compatible stimulation device such that compatible device emulates, in at least one respect, the depictions of sexual activity shown in the video.
2. The system ofclaim 1, further comprising a video analysis engine comprising at least a plurality of programming instructions stored in the memory of, and operating on at least one processor of, the one or more networked computing devices, wherein the plurality of programming instructions, when operating on the at least one processor, cause the one or more networked computing devices to:
input a video containing depictions of sexual activity;
parse the video into at least one component corresponding to one or more of the movements corresponding to the depictions of sexual activity shown in the video;
generate device control data corresponding to the parsed video content; and
store the parsed video content and associated device control data or transmit the parsed video content and associated device control data to a device controller.
3. The system ofclaim 1, further comprising a crowdsourcing portal comprising at least a plurality of programming instructions stored in the memory of, and operating on at least one processor of, the one or more networked computing devices, wherein the plurality of programming instructions, when operating on the processor, cause the computing device the one or more networked computing devices to provide users with compensation for submitting content to the community portal.
4. The system ofclaim 3, further causing the computing device to allow users to establish an online store in the system for sale of content to other users.
5. The system ofclaim 1, further comprising a video selection engine comprising at least a plurality of programming instructions stored in the memory of, and operating on at least one processor of, the one or more networked computing devices, wherein the plurality of programming instructions, when operating on the processor, the one or more networked computing devices to:
allow users to enter preferences as to video content;
search local storage or the internet for videos containing content related to the user's preferences; and
if the videos are not already tagged, allow users to tag the video content; and
use the video content to control a compatible sexual stimulation device.
6. The system ofclaim 1, further comprising automated face-blurring technology.
7. The system ofclaim 1, further comprising account anonymization.
8. The system ofclaim 1, further comprising data encryption technology.
9. The system ofclaim 1, further comprising a trust rating system for interactions with other users.
10. A method for creating and sharing synchronized video control data for sexual stimulation devices, comprising the steps of:
allowing a plurality of users to create user accounts;
allowing the users to choose video content containing depictions of sexual activity from their user accounts;
allowing a user to tag the video content with at least one component corresponding to the depictions of sexual activity shown in the video by manually moving an on-screen object in a synchronized movement with one or more of the movements corresponding to the depictions of sexual activity shown in the video;
generating device control data corresponding to the tagged video content;
allowing users to share the tagged video content and its associated device control data to a community portal;
allowing a plurality of users to access and share videos and associated device control data;
storing the tagged video content and associated device control data; and
transmitting the device control data, synchronized with the video, to a compatible stimulation device such that compatible device emulates, in at least one respect, the depictions of sexual activity shown in the video.
11. The method ofclaim 10, further comprising the steps of:
inputting a video containing depictions of sexual activity;
parsing the video into at least one component corresponding to one or more of the movements corresponding to the depictions of sexual activity shown in the video; and
generating device control data corresponding to the parsed video content; and
storing the parsed video content and associated device control data or transmit the parsed video content and associated device control data to a device controller.
12. The method ofclaim 11, further comprising the steps of:
providing, using a crowdsourcing portal comprising at least a plurality of programming instructions stored in the memory of, and operating on at least one processor of, one or more networked computing devices, users with compensation for submitting content to the community portal.
13. The method ofclaim 12, further comprising the step of:
allowing users to establish an online store in the system for sale of content to other users.
14. The method ofclaim 10, further comprising the steps of:
receiving, at a video selection engine comprising at least a plurality of programming instructions stored in the memory of, and operating on at least one processor of, one or more networked computing devices, user preferences regarding video content;
searching local storage or the internet for videos containing content related to the user's preferences;
if the videos are not already tagged, allow users to tag the video content; and
use the video content to control a compatible sexual stimulation device.
15. The method ofclaim 10, further comprising the step of applying automated face blurring.
16. The method ofclaim 10, further comprising the step of applying account anonymization.
17. The method ofclaim 10, further comprising the step of applying data encryption.
18. The method ofclaim 10, further comprising the step of utilizing a trust rating system for user accounts.
US16/214,0302018-09-242018-12-07Synchronized video control system for sexual stimulation devicesActiveUS10638174B2 (en)

Priority Applications (22)

Application NumberPriority DateFiling DateTitle
US16/214,030US10638174B2 (en)2018-09-242018-12-07Synchronized video control system for sexual stimulation devices
US16/861,014US11185465B2 (en)2018-09-242020-04-28Automated generation of control signals for sexual stimulation devices
US16/942,751US11197074B2 (en)2018-09-242020-07-29Synchronized video annotation and control system for sexual stimulation devices
US17/534,155US11478398B2 (en)2018-09-242021-11-23Automated generation of control signals for sexual stimulation devices
US17/737,968US11590052B2 (en)2018-09-242022-05-05Automated generation of control signals for sexual stimulation devices
US17/737,974US11607366B2 (en)2018-09-242022-05-05Automated generation of initial stimulation profile for sexual stimulation devices
US17/839,426US11596573B2 (en)2018-09-242022-06-13Control of sexual stimulation devices using electroencephalography
US17/853,316US11540971B2 (en)2018-09-242022-06-29Voice-based control of sexual stimulation devices
US17/857,423US20220331196A1 (en)2018-09-242022-07-05Biofeedback-based control of sexual stimulation devices
US17/857,983US11771618B2 (en)2018-09-242022-07-05Adaptive speech and biofeedback control of sexual stimulation devices
US17/970,299US11717467B2 (en)2018-09-242022-10-20Automated generation of control signals for sexual stimulation devices
US18/092,438US20230147864A1 (en)2018-09-242023-01-02Voice-based control of sexual stimulation devices
US18/178,549US11779512B2 (en)2018-09-242023-03-06Control of sexual stimulation devices using electroencephalography
US18/183,948US11766380B2 (en)2018-09-242023-03-14Automated generation of control signals for sexual stimulation devices
US18/185,284US11759387B2 (en)2018-09-242023-03-16Voice-based control of sexual stimulation devices
US18/324,983US12011407B2 (en)2018-09-242023-05-28Automated generation of control signals for sexual stimulation devices
US18/365,158US12213930B2 (en)2018-09-242023-08-03Adaptive speech and biofeedback control of sexual stimulation devices
US18/461,413US20240009072A1 (en)2018-09-242023-09-05Automated generation of control signals for sexual stimulation devices
US18/492,629US20240050306A1 (en)2018-09-242023-10-23Automated generation of control signals for sexual stimulation devices
US18/666,683US20240315917A1 (en)2018-09-242024-05-16Automated generation of control signals for sexual stimulation devices
US18/914,015US20250032350A1 (en)2018-09-242024-10-11Adaptive Speech And Biofeedback Control Of Sexual Stimulation Devices
US18/913,973US20250032349A1 (en)2018-09-242024-10-11Voice-Based Control Of Sexual Stimulation Devices

Applications Claiming Priority (2)

Application NumberPriority DateFiling DateTitle
US16/139,550US10576013B1 (en)2018-09-242018-09-24Synchronized video control system for sexual stimulation devices
US16/214,030US10638174B2 (en)2018-09-242018-12-07Synchronized video control system for sexual stimulation devices

Related Parent Applications (2)

Application NumberTitlePriority DateFiling Date
US16/139,550Continuation-In-PartUS10576013B1 (en)2018-09-242018-09-24Synchronized video control system for sexual stimulation devices
US16/139,550ContinuationUS10576013B1 (en)2018-09-242018-09-24Synchronized video control system for sexual stimulation devices

Related Child Applications (2)

Application NumberTitlePriority DateFiling Date
US16/861,014Continuation-In-PartUS11185465B2 (en)2018-09-242020-04-28Automated generation of control signals for sexual stimulation devices
US16/861,014ContinuationUS11185465B2 (en)2018-09-242020-04-28Automated generation of control signals for sexual stimulation devices

Publications (2)

Publication NumberPublication Date
US20200099970A1 US20200099970A1 (en)2020-03-26
US10638174B2true US10638174B2 (en)2020-04-28

Family

ID=69885158

Family Applications (1)

Application NumberTitlePriority DateFiling Date
US16/214,030ActiveUS10638174B2 (en)2018-09-242018-12-07Synchronized video control system for sexual stimulation devices

Country Status (1)

CountryLink
US (1)US10638174B2 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20220323292A1 (en)*2018-04-102022-10-13Brian SloanControl of sexual stimulation devices using motion-sensing controllers

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
WO2022026646A1 (en)*2020-07-292022-02-03Brian SloanSynchronized video annotation and control system for sexual stimulation devices
WO2023056623A1 (en)*2021-10-092023-04-13Li SiranSexual stimulation method and sexual stimulation apparatus
US11992450B2 (en)*2022-07-192024-05-28AI Video Analysis B.V.Systems and methods for driving an electronic device based on a video feed
US12167988B2 (en)2022-07-192024-12-17AI Video Analysis B.V.Systems and methods for driving an electronic device based on a video feed

Citations (23)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20030108240A1 (en)*2001-12-062003-06-12Koninklijke Philips Electronics N.V.Method and apparatus for automatic face blurring
US20030162595A1 (en)*2002-02-122003-08-28Razz SerbanescuMethod and apparatus for converting sense-preceived thoughts and actions into physical sensory stimulation
US20060270897A1 (en)*2005-05-272006-11-30Homer Gregg SSmart Sex Toys
US20080228580A1 (en)*2007-03-122008-09-18Mynewpedia Corp.Method and system for compensating online content contributors and editors
US20090096632A1 (en)*2007-10-162009-04-16Immersion CorporationSynchronization of haptic effect data in a media stream
US20100041944A1 (en)*2008-08-112010-02-18David LevySexual stimulation devices and methods
US20110133910A1 (en)*2008-10-102011-06-09Internet Services LlcSystem and method for transmitting haptic data in conjunction with media data
US20110321072A1 (en)*2010-06-292011-12-29Google Inc.Self-Service Channel Marketplace
US20120102387A1 (en)*2008-02-192012-04-26Google Inc.Annotating Video Intervals
US20120304216A1 (en)*2011-05-252012-11-29Echostar Technologies L.L.C.Apparatus, systems and methods for presentation management of erotica-related media content
US20130038792A1 (en)*2008-10-102013-02-14Internet Services, LlcSystem and method for synchronization of haptic data and media data
US20130198625A1 (en)*2012-01-262013-08-01Thomas G AndersonSystem For Generating Haptic Feedback and Receiving User Inputs
US20130204957A1 (en)*2009-04-022013-08-08Clear Channel Management Services, Inc.Method, System, and Graphic User Interface for Enabling a Customer to Access a Media File
US20130227410A1 (en)*2011-12-212013-08-29Qualcomm IncorporatedUsing haptic technologies to provide enhanced media experiences
US20130311881A1 (en)*2012-05-162013-11-21Immersion CorporationSystems and Methods for Haptically Enabled Metadata
US8863295B1 (en)2011-09-232014-10-14Google Inc.Linking video sharing accounts with content delivery accounts
US20150262376A1 (en)*2013-03-152015-09-17Immersion CorporationMethod and apparatus to generate haptic feedback from video content analysis
US20150328082A1 (en)*2014-05-162015-11-19HDFEEL Corp.Interactive Entertainment System Having Sensory Feedback
US9396180B1 (en)*2013-01-292016-07-19Amazon Technologies, Inc.System and method for analyzing video content and presenting information corresponding to video content to users
WO2016201576A1 (en)2015-06-162016-12-22Standard Innovation CorporationSensor acquisition and analytics platform for enhancing interaction with adult devices
US20170366858A1 (en)*2016-06-202017-12-21Shenzhen Love Sense Technology Co., Ltd.System and method for composing function programming for adult toy operation in synchronization with video playback
US20180012032A1 (en)*2014-10-232018-01-11Pageproof.Com LimitedEncrypted collaboration system and method
US10091462B2 (en)2006-12-052018-10-02Crackle, Inc.Video sharing platform providing for posting content to other websites

Patent Citations (23)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20030108240A1 (en)*2001-12-062003-06-12Koninklijke Philips Electronics N.V.Method and apparatus for automatic face blurring
US20030162595A1 (en)*2002-02-122003-08-28Razz SerbanescuMethod and apparatus for converting sense-preceived thoughts and actions into physical sensory stimulation
US20060270897A1 (en)*2005-05-272006-11-30Homer Gregg SSmart Sex Toys
US10091462B2 (en)2006-12-052018-10-02Crackle, Inc.Video sharing platform providing for posting content to other websites
US20080228580A1 (en)*2007-03-122008-09-18Mynewpedia Corp.Method and system for compensating online content contributors and editors
US20090096632A1 (en)*2007-10-162009-04-16Immersion CorporationSynchronization of haptic effect data in a media stream
US20120102387A1 (en)*2008-02-192012-04-26Google Inc.Annotating Video Intervals
US20100041944A1 (en)*2008-08-112010-02-18David LevySexual stimulation devices and methods
US20130038792A1 (en)*2008-10-102013-02-14Internet Services, LlcSystem and method for synchronization of haptic data and media data
US20110133910A1 (en)*2008-10-102011-06-09Internet Services LlcSystem and method for transmitting haptic data in conjunction with media data
US20130204957A1 (en)*2009-04-022013-08-08Clear Channel Management Services, Inc.Method, System, and Graphic User Interface for Enabling a Customer to Access a Media File
US20110321072A1 (en)*2010-06-292011-12-29Google Inc.Self-Service Channel Marketplace
US20120304216A1 (en)*2011-05-252012-11-29Echostar Technologies L.L.C.Apparatus, systems and methods for presentation management of erotica-related media content
US8863295B1 (en)2011-09-232014-10-14Google Inc.Linking video sharing accounts with content delivery accounts
US20130227410A1 (en)*2011-12-212013-08-29Qualcomm IncorporatedUsing haptic technologies to provide enhanced media experiences
US20130198625A1 (en)*2012-01-262013-08-01Thomas G AndersonSystem For Generating Haptic Feedback and Receiving User Inputs
US20130311881A1 (en)*2012-05-162013-11-21Immersion CorporationSystems and Methods for Haptically Enabled Metadata
US9396180B1 (en)*2013-01-292016-07-19Amazon Technologies, Inc.System and method for analyzing video content and presenting information corresponding to video content to users
US20150262376A1 (en)*2013-03-152015-09-17Immersion CorporationMethod and apparatus to generate haptic feedback from video content analysis
US20150328082A1 (en)*2014-05-162015-11-19HDFEEL Corp.Interactive Entertainment System Having Sensory Feedback
US20180012032A1 (en)*2014-10-232018-01-11Pageproof.Com LimitedEncrypted collaboration system and method
WO2016201576A1 (en)2015-06-162016-12-22Standard Innovation CorporationSensor acquisition and analytics platform for enhancing interaction with adult devices
US20170366858A1 (en)*2016-06-202017-12-21Shenzhen Love Sense Technology Co., Ltd.System and method for composing function programming for adult toy operation in synchronization with video playback

Cited By (2)

* Cited by examiner, † Cited by third party
Publication numberPriority datePublication dateAssigneeTitle
US20220323292A1 (en)*2018-04-102022-10-13Brian SloanControl of sexual stimulation devices using motion-sensing controllers
US12011408B2 (en)*2018-04-102024-06-18Brian SloanControl of sexual stimulation devices using motion-sensing controllers

Also Published As

Publication numberPublication date
US20200099970A1 (en)2020-03-26

Similar Documents

PublicationPublication DateTitle
US11197074B2 (en)Synchronized video annotation and control system for sexual stimulation devices
US10638174B2 (en)Synchronized video control system for sexual stimulation devices
US10576013B1 (en)Synchronized video control system for sexual stimulation devices
US11694726B2 (en)Automatic trailer detection in multimedia content
US11080616B2 (en)Artificial intelligence model and data collection/development platform
US20250005449A1 (en)Systems and methods for customizing user interfaces using artificial intelligence
US11681943B2 (en)Artificial intelligence development via user-selectable/connectable model representations
US10397359B2 (en)Streaming media cache for media streaming service
US11478398B2 (en)Automated generation of control signals for sexual stimulation devices
US11040168B2 (en)Cognitive stress management by augmented reality (AR) customization
CN106789543A (en)The method and apparatus that facial expression image sends are realized in session
KR20160015319A (en)Generating a feed of content items associated with a topic from multiple content sources
US20180063064A1 (en)Modifying a mood through selective feeding of content
US11204955B2 (en)Digital content delivery based on predicted effect
WO2019227630A1 (en)Song recommending method and apparatus
US20150052440A1 (en)Real-time management of presentation delivery
US10790055B2 (en)Appetite improvement system through memory association
US20250005444A1 (en)Systems and methods for customizing user interfaces using artificial intelligence
US12400029B2 (en)System and method for securing and facilitating access to a digital legacy
JP7691590B2 (en) Method, content manager, and program for automatically generating extensions to AV content
US20170206586A1 (en)Dynamically Customizing User Quotas
CN112269942B (en)Method, device and system for recommending object and electronic equipment
WO2022026646A1 (en)Synchronized video annotation and control system for sexual stimulation devices
Ribezzo et al.TAPAS-360: A Tool for the Design and Experimental Evaluation of 360 Video Streaming Systems
US11630865B2 (en)User reaction based information options

Legal Events

DateCodeTitleDescription
FEPPFee payment procedure

Free format text:ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

FEPPFee payment procedure

Free format text:ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

STCFInformation on status: patent grant

Free format text:PATENTED CASE

MAFPMaintenance fee payment

Free format text:PAYMENT OF MAINTENANCE FEE, 4TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2551); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

Year of fee payment:4


[8]ページ先頭

©2009-2025 Movatter.jp