An augmented reality (AR) eyewear device has a lens system which includes an optical screening mechanism that enables switching the lens system between a conventional see-through state and an opaque state in which the lens system screens or functionally blocks out the wearer's view of the external environment. Such a screening mechanism allows for expanded use cases of the AR glasses compared to conventional devices, e.g.: as a sleep mask; to view displayed content like movies or sports events against a visually non-distracting background instead of against the external environment; and/or to enable VR functionality.
A system for hand tracking for an Augmented Reality (AR) system. The AR system uses a camera of the AR system to capture tracking video frame data of a hand of a user of the AR system. The AR system generates a skeletal model based on the tracking video frame data and determines a location of the hand of the user based on the skeletal model. The AR system causes a steerable camera of the AR system to focus on the hand of the user.
A drone system is configured to capture an audio stream that includes voice commands from an operator, to process the audio stream for identification of the voice commands, and to perform operations based on the identified voice commands. The drone system can identify a particular voice stream in the audio stream as an operator voice, and perform the command recognition with respect to the operator voice to the exclusion of other voice streams present in the audio stream. The drone can include a directional camera that is automatically and continuously focused on the operator to capture a video stream usable in disambiguation of different voice streams captured by the drone.
Embodiments described herein relate to an augmented expression system to generate and cause display of a specially configured interface to present an augmented reality perspective. The augmented expression system receives image and video data of a user and tracks facial landmarks of the user based on the image and video data, in real-time to generate and present a 3-dimensional (3D) bitmoji of the user.
Aspects of the present disclosure involve a system comprising a computer-readable storage medium storing a program and method for providing bot participants for virtual conferencing. The program and method provide, in association with designing a virtual space, a first interface for configuring plural participant video elements, each being assignable to a respective participant; receive, via the first interface, an indication of user input for setting first properties for the plural participant video elements; provide a second interface for configuring a bot participant for simulating an actual participant in association with a participant video element of the plural participant video elements; receive, via the second interface, an indication of second user input for setting second properties for the bot participant; and provide, in association with designing the virtual space, display of the virtual space based on the first and second properties, the bot participant being assigned to the participant video element.
G06F 3/04847 - Interaction techniques to control parameter settings, e.g. interaction with sliders or dials
G06F 3/04815 - Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
G06T 5/20 - Image enhancement or restoration by the use of local operators
H04L 51/02 - User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail using automatic reactions or user delegation, e.g. automatic replies or chatbot-generated messages
H04L 65/403 - Arrangements for multi-party communication, e.g. for conferences
6.
FACIAL SYNTHESIS IN AUGMENTED REALITY CONTENT FOR ONLINE COMMUNITIES
The subject technology captures first image data by a computing device, the first image data comprising a target face of a target actor and facial expressions of the target actor, the facial expressions including lip movements. The subject technology generates, based at least in part on frames of a source media content, sets of source pose parameters. The subject technology receives a selection of a particular facial expression from a set of facial expressions. The subject technology generates, based at least in part on sets of source pose parameters and the selection of the particular facial expression, an output media content. The subject technology provides augmented reality content based at least in part on the output media content for display on the computing device.
Aspects of the present disclosure involve a system comprising a computer-readable storage medium storing a program and method for dynamically assigning participant video feeds within a virtual conferencing system. The program and method provide, in association with designing a virtual space for virtual conferencing, an interface for configuring a set of rooms, each room being associated with a different number of participant video elements assignable to respective participant video feeds; receive, via the interface, an indication of user input for setting properties for the set of rooms; determine, in association with virtual conferencing, a first number of participants for a room; select a first room corresponding to the first number of participants; provide display of the first room; and assign, for each of the first number of participants, a participant video feed corresponding to the participant with a respective participant video element in the first room.
G06F 3/04815 - Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
G06F 3/0482 - Interaction with lists of selectable items, e.g. menus
G06F 3/0484 - Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
8.
SOFTWARE APPLICATION MANAGER FOR MESSAGING APPLICATIONS
Among other things, embodiments of the present disclosure improve the functionality of electronic messaging systems by enabling users in an electronic chat conversation to run applications together. In some embodiments, when one user in a chat launches an application, an icon or other visual representation of the application appears in a portion of the chat window (e.g., in a “chat dock”) for other users in the chat to access.
H04L 65/401 - Support for services or applications wherein the services involve a main real-time session and one or more additional parallel real-time or time sensitive sessions, e.g. white board sharing or spawning of a subconference
Devices, media, and methods are presented for an immersive augmented reality (AR) experience using an eyewear device with spatial audio. The eyewear device has a processor, a memory, an image sensor, and a speaker system. The eyewear device captures image information for an environment surrounding the device and identifies an object location within the same environment. The eyewear device then associates a virtual object with the identified object location. The eyewear device monitors the position of the device with respect to the virtual object and presents audio signals to alert the user that the identified object is in the environment.
Systems and methods are provided for performing operations comprising: capturing, by an electronic mirroring device, a video feed received from a camera of the electronic mirroring device, the video feed depicting a user; displaying, by one or more processors of the electronic mirroring device, one or more menu options on the video feed that depicts the user, the one or more menu options relating to a first level in a hierarchy of levels; detecting a gesture performed by the user in the video feed; and in response to detecting the gesture, displaying a set of options related to a given option of the one or more menu options, the set of options relating to a second level in the hierarchy of levels.
A method of generating an image for use in a conversation taking place in a messaging application is disclosed. Conversation input text is received from a user of a portable device that includes a display. Model input text is generated from the conversation input text, which is processed with a text-to-image model to generate an image based on the model input text. The coordinates of a face in the image are determined, and the face of the user or another person is added to the image at the location. The final image is displayed on the portable device, and user input is received to transmit the image to a remote recipient.
Systems and methods are provided for retrieving first query result data associated with a first user account and rendering the first query result data into a first result item, generating a shareable search result stream comprising the first result item associated with the first user account, retrieving second query result data associated with a second user account and rendering the second query result data into a second result item, adding the second result item to the shareable search result stream associated with the first user account, and providing the sharable search result stream comprising the first result item and the second result item to a first computing device associated with the first user account and a second computing device associated with the second user account.
Methods and systems are disclosed for performing operations for transferring garments from one real-world object to another in real time. The operations comprise receiving a first video that includes a depiction of a first person wearing a first upper-body garment in a first pose and obtaining a second video that includes a depiction of a second person wearing a second upper-body garment in a second pose. A pose of the second person depicted in the second video is modified to match the first pose of the first person depicted in the first video. The operations comprise generating an upper-body segmentation of the second upper-body garment which the second person is wearing in the second video in the modified pose and replacing the first upper-body garment worn by the first person in the first video with the second upper-body garment based on the upper-body segmentation.
G06T 7/70 - Determining position or orientation of objects or cameras
G06T 19/20 - Editing of 3D images, e.g. changing shapes or colours, aligning objects or positioning parts
14.
DEVICE AND METHOD FOR COMPENSATING EFFECTS OF PANTOSCOPIC TILT OR WRAP/SWEEP TILT ON AN IMAGE PRESENTED ON AN AUGMENTED REALITY OR VIRTUAL REALITY DISPLAY
An optical device is disclosed for use in an augmented reality or virtual reality display, comprising a waveguide (12; 22; 32) and an input diffractive optical element (H0; H3; 34) positioned in or on the waveguide, configured to receive light from a projector and couple it into the waveguide so that it is captured within the waveguide under total internal reflection. The input diffractive optical element has an input grating vector (G0; Gig) in the plane of the waveguide. The device includes a first diffractive optical element (H1; H4) and a second diffractive optical element (H2; H5) having first and second grating vectors (G2, G3; GV1, GV2) respectively in the plane of the waveguide, wherein the first diffractive optical element is configured to receive light from the input diffractive optical element and to couple it towards the second diffractive optical element, and wherein the second diffractive optical element is configured to receive light from the first diffractive optical element and to couple it out of the waveguide towards a viewer. The input grating vector, the first grating vector and the second grating vector have different respective magnitudes, and wherein a vector addition of the input grating vector, the first grating vector and the second grating vector sums to zero.
Aspects of the present disclosure involve a system comprising a computer-readable storage medium storing a program and method for configuring a three-dimensional (3D) model within a virtual conferencing system. The program and method provide, in association with designing a room for virtual conferencing, an interface for configuring a 3D model; receiving, via the interface, an indication of user input for setting properties for the 3D model, the properties specifying image data for projecting onto the 3D model; and in association with virtual conferencing, providing display of the room based on the properties for the 3D model, and causing the image data to be projected onto the 3D model within the room.
G06F 3/04815 - Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
G06F 3/04845 - Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range for image manipulation, e.g. dragging, rotation, expansion or change of colour
G06F 3/04847 - Interaction techniques to control parameter settings, e.g. interaction with sliders or dials
G06T 19/00 - Manipulating 3D models or images for computer graphics
G06T 19/20 - Editing of 3D images, e.g. changing shapes or colours, aligning objects or positioning parts
A method and a system include receiving a request from a client device to view a media content item, determining at least one comment associated with a respective user profile from a set of connected profiles, generating a summary comments selectable item based at least in part on the respective user profile, causing a display of playback of the media content item and the summary comments selectable item in response to the request to view the media content item, and during the playback of the media content item at the particular time, causing a display of at least one comment.
H04L 51/52 - User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail for supporting social networking services
G06F 3/04817 - Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance using icons
Methods and devices for wired charging and communication with a wearable device are described. In one embodiment, a symmetrical contact interface comprises a first contact pad and a second contact pad, and particular wired circuitry is coupled to the first and second contact pads to enable charging as well as receive and transmit communications via the contact pads as part of various device states.
H01L 27/02 - Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including integrated passive circuit elements with at least one potential-jump barrier or surface barrier
H01R 13/62 - Means for facilitating engagement or disengagement of coupling parts or for holding them in engagement
H02J 7/04 - Regulation of the charging current or voltage
H02J 7/34 - Parallel operation in networks using both storage and other dc sources, e.g. providing buffering
H03K 19/0185 - Coupling arrangements; Interface arrangements using field-effect transistors only
H04B 3/56 - Circuits for coupling, blocking, or by-passing of signals
18.
SELECTING ITEMS DISPLAYED BY A HEAD-WORN DISPLAY DEVICE
Disclosed is a method of receiving and processing content-sending inputs received by a head-worn device system including one or more display devices, one or more cameras and a vertically-arranged touchpad. The method includes displaying a content item on the one or more display devices, receiving a touch input on the touchpad corresponding to a send instruction, displaying a carousel of potential recipients, receiving a horizontal touch input on the touchpad, scrolling the carousel left or right on the one or more display devices in response to the horizontal touch input, receiving a tap touch input on the touchpad to select a particular recipient, receiving a further touch input, and in response to the further touch input, transmitting the content item to the selected recipient.
G06F 3/04815 - Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
G06F 3/0488 - Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
G06T 19/00 - Manipulating 3D models or images for computer graphics
Aspects of the present disclosure involve a system and a method for performing operations comprising: receiving, by a messaging application implemented on a client device, input that selects a sound option to add sound to one or more images; in response to receiving the input, presenting a sound editing user interface element that visually indicates a played portion of the sound and separately visually indicates an un-played portion of the sound; receiving an interaction with the sound editing user interface element to modify a start point of the sound; embedding a graphical element representing the sound in the one or more images; playing, by the messaging application, the sound associated with the graphical element starting from the start point together with displaying the one or more images.
A gesture-based wake process for an AR system is described herein. The AR system places a hand-tracking input pipeline of the AR system in a suspended mode. A camera component of the hand-tracking input pipeline detects a possible visual wake command being made by a user of the AR system. On the basis of detecting the possible visual wake command, the AR system wakes the hand-tracking input pipeline and places the camera component in a fully operational mode. If the AR system, using the hand-tracking input pipeline, verifies the possible visual wake command as an actual wake command, the AR system initiates execution of an AR application.
A system includes a communication module that receives a request to post content to an event gallery associated with an event. The request in turn includes geo-location data for a device sending the content, and identification data identifying the device or a user of the device. The system further has an event gallery module to perform a first authorization operation that includes determining that the geo-location data corresponds to a geo-location fence associated with an event. The event gallery module also performs a second authorization operation that includes using the identification data to verify an attribute of the user. Finally, based on the first and second authorization operations, the event gallery module may selectively authorize the device to post the content to the event gallery.
H04L 51/222 - Monitoring or handling of messages using geographical location information, e.g. messages transmitted or received in proximity of a certain spot or area
H04L 51/52 - User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail for supporting social networking services
H04W 4/02 - Services making use of location information
H04W 4/021 - Services related to particular areas, e.g. point of interest [POI] services, venue services or geofences
H04W 4/029 - Location-based management or tracking services
H04W 4/18 - Information format or content conversion, e.g. adaptation by the network of the transmitted or received information for the purpose of wireless delivery to users or terminals
Systems and methods herein describe a method for capturing a video in real-time by an image capture device. The system provides a plurality of visual pose hints, identifies first pose information in the video while capturing the video, applies a first series of virtual effects to the video, identifies second pose information, and applies a second series of virtual effects to the video, the second series of virtual effects based on the first series of virtual effects.
A push notification mechanism at a mobile user device provides for automated limiting of the rate of production of push notification alerts (such as an audible alert or a vibratory alert) and/or push notifications responsive to the occurrence of chat events relevant to a chat application hosted by the user device. Some chat events automatically trigger suppression periods during which push notification alerts are prevented for subsequent chat events that satisfy predefined suppression criteria. Such push notification and/or alert limiting can be performed separately for separate users, chat groups, and/or chat event types.
H04L 51/04 - Real-time or near real-time messaging, e.g. instant messaging [IM]
H04L 51/224 - Monitoring or handling of messages providing notification on incoming messages, e.g. pushed notifications of received messages
H04L 51/52 - User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail for supporting social networking services
Various embodiments include systems, methods, and non-transitory computer-readable media for sharing and managing media galleries. Consistent with these embodiments, a method includes receiving a request from a first device to share a media gallery that includes a user avatar; generating metadata associated with the media gallery; generating a message associated with the media gallery, the message at least including the media gallery identifier and the identifier of the user avatar; and transmitting the message to a second device of the recipient user.
G06F 3/0482 - Interaction with lists of selectable items, e.g. menus
G06F 3/0484 - Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
H04L 67/146 - Markers for unambiguous identification of a particular session, e.g. session cookie or URL-encoding
25.
GRAPHICAL ASSISTANCE WITH TASKS USING AN AR WEARABLE DEVICE
Systems, methods, and computer readable media for graphical assistance with tasks using an augmented reality (AR) wearable devices are disclosed. Embodiments capture an image of a first user view of a real-world scene and access indications of surfaces and locations of the surfaces detected in the image. The AR wearable device displays indications of the surfaces on a display of the AR wearable device where the locations of the indications are based on the locations of the surfaces and a second user view of the real-world scene. The locations of the surfaces are indicated with 3D world coordinates. The user views are determined based on a location of the user. The AR wearable device enables a user to add graphics to the surfaces and select tasks to perform. Tools such as a bubble level or a measuring tool are available for the user to utilize to perform the task.
A content controller system to render frames on demand comprises a rendering server system that includes a plurality of graphics processing units (GPUs). The GPUs in the rendering server system render a set of media content item segments using a media content identification and a main user identification. Rendering the set of media content item segments includes retrieving metadata from a metadata database associated with the media content identification, rendering the set of media content item segments using the metadata, generating a main user avatar based on the main user identification, and incorporating the main user avatar into the set of media content item segments. The rendering server system then uploads the set of media content item segments to a segment database; and updates segment states in a segment state database to indicate that the set of media content item segments are available. Other embodiments are disclosed herein.
H04N 21/262 - Content or additional data distribution scheduling, e.g. sending additional data at off-peak times, updating software modules, calculating the carousel transmission frequency, delaying a video stream transmission or generating play-lists
G06T 1/20 - Processor architectures; Processor configuration, e.g. pipelining
H04N 21/234 - Processing of video elementary streams, e.g. splicing of video streams or manipulating MPEG-4 scene graphs
H04N 21/235 - Processing of additional data, e.g. scrambling of additional data or processing content descriptors
H04N 21/239 - Interfacing the upstream path of the transmission network, e.g. prioritizing client requests
H04N 21/258 - Client or end-user data management, e.g. managing client capabilities, user preferences or demographics or processing of multiple end-users preferences to derive collaborative data
H04N 21/84 - Generation or processing of descriptive data, e.g. content descriptors
Eyewear devices including a tether and methods for identifying proper installation of the tether are disclosed. An eyewear device includes transmission lines extending through the temples to electrical and electronic components positioned adjacent to edges of a frame. A tether is attached to the temples to enable power and communication flow between the electrical and electronic components rather than through the frame. Proper installation is identified based on communications passing between the electrical and electronic components via the tether.
Methods and systems are disclosed for performing operations for controlling brightness in an AR device. The operations comprise displaying an image on an eyewear device worn by a user; detecting a gaze direction of a pupil of the user; identifying a first region of the image that corresponds to the gaze direction of the pupil; and modifying a brightness level or value of pixels in the image based on the gaze direction such that pixels in the first region of the image are set to a first brightness value and pixels in a second region of the image are set to a second brightness value that is lower than the first brightness value.
The subject technology detects from a set of frames, a first gesture, the first gesture corresponding to a pinch gesture. The subject technology detects a first location and a first position of a first representation of a first finger from the first gesture and a second location and a second position of a second representation of a second finger from the first gesture. The subject technology detects a first collision event corresponding to a first collider and a second collider intersecting with a third collider of a first virtual object. The subject technology detects a first change in the first location and the first position and a second change in the second location and the second position. The subject technology modifies the first virtual object to include an additional augmented reality content based at least in part on the first change and the second change.
Systems and methods are provided for clustering videos. The system accesses a plurality of content items, the plurality of content items comprising a first set of RGB video frames and a second set of optical flow frames corresponding to the first set of RGB video frames. The system processes the first set of RGB video frames by a first machine learning model to generate a first optimal assignment for the first set of RGB video frames, the first optimal assignment representing initial clustering of the first set of RGB video frames. The system generates an updated first optimal assignment for the first set of RGB video frames based on the first optimal assignment for the first set of RGB video frames and a second optimal assignment of the second set of optical flow frames, the second optimal assignment representing initial clustering of the second set of optical flow frames.
G06V 10/762 - Arrangements for image or video recognition or understanding using pattern recognition or machine learning using clustering, e.g. of similar faces in social networks
G06T 5/50 - Image enhancement or restoration by the use of more than one image, e.g. averaging, subtraction
G06V 10/82 - Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
31.
VIRTUAL OBJECT MANIPULATION WITH GESTURES IN A MESSAGING SYSTEM
The subject technology detects a first gesture and a second gesture, each gesture corresponding to an open trigger finger gesture. The subject technology detects a third gesture and a fourth gesture, each gesture corresponding to a closed trigger finger gesture. The subject technology, selects a first virtual object in a first scene. The subject technology detects a first location and a first position of a first representation of a first finger from the third gesture and a second location and a second position of a second representation of a second finger from the fourth gesture. The subject technology detects a first change in the first location and the first position and a second change in the second location and the second position. The subject technology modifies a set of dimensions of the first virtual object to a different set of dimensions.
The subject technology detects a first location and a first position of a first representation of a first finger and a second location and a second position of a second representation of a second finger. The subject technology detects a first particular location and a first particular position of a first particular representation of a first particular finger and a second particular location and a second particular position of a second particular representation of a second particular finger. The subject technology detects a first change in the first location and the first position and a second change in the second location and the second position. The subject technology detects a first particular change in the first particular location and the first particular position and a second particular change in the second particular location and the second particular position. The subject technology generates a set of virtual objects.
Methods and systems are disclosed for performing real-time deforming operations. The system receives an image that includes a depiction of a real-world object. The system applies a machine learning model to the image to generate a warping field and segmentation mask, the machine learning model trained to establish a relationship between a plurality of training images depicting real-world objects and corresponding ground-truth warping fields and segmentation masks associated with a target shape. The system applies the generated warping field and segmentation mask to the image to warp the real-world object depicted in the image to the target shape.
In various embodiments, boundaries of geo-fences can be made mutable based on principles described herein. The term “mutable” refers to the ability of a thing (in this case, the boundary of a geo-fence) to change and adjust. In a typical embodiment, a mutable geo-fence system is configured to generate and monitor a geo-fence that encompasses a region, in order to dynamically vary the boundary of the geo-fence based on a number of boundary variables. The term “geo-fence” as used herein describes a virtual perimeter (e.g., a boundary) for a real-world geographic area. A geo-fence could be a radius around a point (e.g., a store), or a set of predefined boundaries. Boundary variables, as used herein, refers to a set of variables utilized by the mutable geo-fence system in determining a location of the boundary of the geo-fence.
Augmented reality guidance for guiding a user through an environment using an eyewear device. The eyewear device includes a display system and a position detection system. A user is guided though an environment by monitoring a current position of the eyewear device within the environment, identifying marker positions within a threshold of the current position, the marker positions defined with respect to the environment and associated with guidance markers, registering the marker positions, generating overlay image including the guidance markers, and presenting the overlay image on a display of the eyewear device.
Aspects of the present disclosure involve a system comprising a computer-readable storage medium storing a program and method for selecting ads for a video. The program and method provide for receiving a request for an ad to insert into a video playing on a client device, the request including a first content identifier that identifies a first type of content included in the video; determining a set of content identifiers associated with the first content identifier, the set of content identifiers identifying second types of content to filter with respect to providing the ad in response to the request; selecting an ad from among plural ads, by filtering ads tagged with a second content identifier included in the set of content identifiers; and providing the selected ad as a response to the request.
H04N 21/44 - Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to MPEG-4 scene graphs
H04N 21/4788 - Supplemental services, e.g. displaying phone caller identification or shopping application communicating with other users, e.g. chatting
A carry case for an electronics-enabled eyewear device, such as smart glasses, has charging contacts that are movable relative to a storage chamber in which the eyewear device is receivable. The charging contacts are connected to a battery carried by the case for charging the eyewear device via contact coupling of the charging contacts to corresponding contact formations on an exterior of the eyewear device. The charging contacts are in some instances mounted on respective flexible walls defining opposite extremities of the storage chamber. The contact formations on the eyewear device are in some instances provided by hinge assemblies that couple respective temples to a frame of the eyewear device.
Methods, systems, and devices are described for predicting a destination of a user and sharing the presumed destination with the other users via a geographically-based graphical user interface. Consistent with some embodiments, an electronic communication containing location information is received from a location sensor coupled to a first client device. A current trajectory of the first user is determined based on the location information. A presumed destination of the first user is determined, by correlating the current trajectory of the first user with historical location information of the first user. A map depicting an icon associated with the presumed destination of the first user is displayed, on a display screen of a second client device of a second user.
An eyewear device is disclosed including an illumination device including illumination sources, each illumination source including a first illuminator, a second illuminator, and a third illuminator, and a spatial light modulator coupled to the illumination device to control when each of the first, second, and third illuminators are on during an illumination frame. The spatial light modulator is adapted to turn on the first illuminator while the second and third illuminators are off, turn on the second illuminator while the first and third illuminators are off, turn on the third illuminator while the first and second illuminators are off during a third time period of the illumination frame, and turn on the first, second and third illuminators during a fourth time period. An illumination method is also disclosed.
Described are systems and methods for generating personalized videos with customized text messages. An example method includes receiving an input text, a video template including a sequence of frame images, and at least one parameter for animation of the input text across the sequence of frame images, generating, based on the input text and the at least one parameter for animation, a configuration file including a text style for the input text for a frame in the sequence of frame images, and rendering, based on the configuration file, an output frame of an output video, where the output frame includes the frame in the sequence of frame images and a layer, and where the layer includes the input text stylized based on the text style. The method further includes providing an option enabling a user to change the at least one parameter for animation.
Provided are systems and methods for customizing modifiable videos. An example method includes analyzing recent messages associated with a user in a multimedia messaging application to determine a context of the recent messages, determining, based on the context, a property of a modifiable feature, selecting, based on the context, a list of relevant modifiable videos from a database configured to store modifiable videos associated with a preset modifiable feature, replacing a property of the preset modifiable feature in relevant modifiable videos of the list of relevant modifiable videos with the property of the modifiable feature, and rendering the list of relevant modifiable videos for viewing by the user, where the rendering includes displaying the modifiable feature in the relevant modifiable videos.
H04M 1/72439 - User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for image or video messaging
G06T 11/60 - Editing figures and text; Combining figures or text
H04L 51/02 - User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail using automatic reactions or user delegation, e.g. automatic replies or chatbot-generated messages
H04L 51/216 - Handling conversation history, e.g. grouping of messages in sessions or threads
H04M 1/72427 - User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality for supporting games or graphical animations
H04M 1/72436 - User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for text messaging, e.g. SMS or e-mail
H04M 1/72442 - User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality for playing music files
Aspects of the present disclosure involve a system comprising a computer-readable storage medium storing a program and a method for performing operations comprising: receiving a plurality of records associated with a first geographical area; identifying a plurality of corrections to a first attribute in the first geographical area in the plurality of records for a particular time period; based on identifying the plurality of corrections to the first attribute, computing a first metric representing a quantity of the plurality of corrections to the first attribute per effort during the particular time period; accumulating a first value representing a total number of errors across a plurality of time periods up to and including the particular time period based on the identified plurality of corrections; and generating a first model that predicts accuracy of the first attribute in the first geographical area based on the metric and the accumulated first value.
Context-sensitive remote controls for use with electronic devices (e.g., eyewear device). The electronic device is configured to perform activities (e.g., email, painting, navigation, gaming). The context-sensitive remote control includes a display having a display area, a display driver coupled to the display, and a transceiver. The remote control additionally includes memory that stores controller layout configurations for display in the display area of the display by the display driver. A processor in the context-sensitive remote control is configured to establish, via the transceiver, communication with an electronic device, detect an activity currently being performed by the electronic device, select one of the controller layout configurations responsive to the detected activity, and present, via the display driver, the selected controller layout configuration in the display area of the display.
A63F 13/22 - Setup operations, e.g. calibration, key configuration or button assignment
A63F 13/211 - Input arrangements for video game devices characterised by their sensors, purposes or types using inertial sensors, e.g. accelerometers or gyroscopes
A63F 13/2145 - Input arrangements for video game devices characterised by their sensors, purposes or types for locating contacts on a surface, e.g. floor mats or touch pads the surface being also a display device, e.g. touch screens
A63F 13/24 - Constructional details thereof, e.g. game controllers with detachable joystick handles
A63F 13/327 - Interconnection arrangements between game servers and game devices; Interconnection arrangements between game devices; Interconnection arrangements between game servers using local area network [LAN] connections using wireless networks, e.g. Wi-Fi or piconet
A63F 13/40 - Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment
A63F 13/92 - Video game devices specially adapted to be hand-held while playing
G06F 3/0346 - Pointing devices displaced or positioned by the user; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
G06F 3/038 - Control and interface arrangements therefor, e.g. drivers or device-embedded control circuitry
G06F 3/04842 - Selection of displayed objects or displayed text elements
G06F 3/04886 - Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures by partitioning the display area of the touch-screen or the surface of the digitising tablet into independently controllable areas, e.g. virtual keyboards or menus
An apparatus to perform fast data access comprises a receiver, a processor, and a memory. The processor receives using the receiver a light signal from a light source. The light signal can be structured to generate a temporal code. The light source is an optical beacon that includes a Light-Emitting Diode (LED). The processor then decodes the light signal to generate a network address, and causes a display of a client device coupled to the apparatus to display information based on the network address. The network address can be a Uniform Resource Locator (URL) address and the information based on the network address includes a webpage associated with the URL. Other embodiments are described herein.
G06K 7/14 - Methods or arrangements for sensing record carriers by corpuscular radiation using light without selection of wavelength, e.g. sensing reflected white light
G06K 7/10 - Methods or arrangements for sensing record carriers by corpuscular radiation
Systems and methods for radial gesture navigation are provided. In example embodiments, user input data is received from a user device. The user input data indicates a continuous physical user interaction associated with a display screen of the user device. An initial point and a current point are detected from the user input data. A radius distance for a circle that includes the current point and is centered about the initial point is determined. An action is selected from among multiple actions based on the radius distance being within a particular range among successive ranges along a straight line that starts at the initial point and extends through the circle. Each range among the successive ranges corresponds to a particular action among the multiple actions. The selected action is performed in response to detecting a completion of the continuous physical user interaction.
Among other things, embodiments of the present disclosure improve the functionality of electronic messaging software and systems by generating customized images with avatars of different users within electronic messages. For example, users of different mobile computing devices can exchange electronic communications with images generated to include avatars representing themselves as well as their friends, colleagues, and other acquaintances.
H04L 51/216 - Handling conversation history, e.g. grouping of messages in sessions or threads
H04L 51/52 - User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail for supporting social networking services
H04L 67/10 - Protocols in which an application is distributed across nodes in the network
The subject technology receives frames of a source media content. The subject technology detects from the frames of the source media content, a first gesture indicating a cut point at a particular frame of the source media content, the cut point associated with a trimming operation to be performed on the source media content. The subject technology selects a starting frame and an ending frame from the frames based at least in part on the cut point at the particular frame. The subject technology performs the trimming operation based on the starting frame and the ending frame. The subject technology generates a second media content using the third set of frames. The subject technology provides for display at least a portion of the third set of frames of the second media content.
The subject technology detects a location and a position of a representation of a finger. The subject technology generates a first virtual object based on the location and the position of the representation of the finger. The subject technology detects a first collision event. The subject technology in response to the first collision event, modifies a set of dimensions of the second virtual object to a second set of dimensions. The subject technology detects a second location and a second position of the representation of the finger. The subject technology detects a second collision event. The subject technology modifies a set of dimensions of the third virtual object to a third set of dimensions. The subject technology renders the third virtual object based on the third set of dimensions within a third scene, the third scene comprising a modified scene from a second scene.
The subject technology receives a set of frames. The subject technology detect a first gesture correspond to an open trigger finger gesture. The subject technology receives a second set of frames. The subject technology detects from the second set of frames, a second gesture correspond to a closed trigger finger gesture. The subject technology detects a location and a position of a representation of a finger from the closed trigger finger gesture. The subject technology generates a first virtual object based at least in part on the location and the position of the representation of the finger. The subject technology renders a movement of the first virtual object along a vector away from the location and the position of the representation of the finger within a first scene. The subject technology provides for display the rendered movement of the first virtual object along the vector within the first scene.
The subject technology detects a location and a position of a representation of a finger in a set of frames captured by a camera of a client device. The subject technology generates a first virtual object based at least in part on the location and the position of the representation of the finger. The subject technology renders the first virtual object within a first scene. The subject technology detects a first collision event corresponding to a first collider of the first virtual object intersecting with a second collider of a second virtual object. The subject technology modifies a set of dimensions of the second virtual object to a second set of dimensions. The subject technology renders the second virtual object based on the second set of dimensions within a second scene. The subject technology provides for display the rendered second virtual object within the second scene.
Systems, methods, and computer readable media are described for remotely changing settings on augmented reality (AR) wearable devices. Embodiments are disclosed that enable a user to change settings of an AR wearable device on a user interface (UI) provided by a host client device that can communicate wirelessly with the AR wearable device. The host client device and AR wearable device provide remote procedure calls (RPCs) and an application program interface (API) to access settings and determine if settings have been changed. The API enables the host client device to determine the settings on the AR wearable device without any prior knowledge of the settings on the AR wearable device. The RPCs and the API enable the host client device to automatically update the settings on the AR wearable device when the user changes the settings on the host client device.
Systems, methods, and computer readable media for selecting a tilt angle of an augmented reality (AR) display of an AR wearable device. Some examples of the present disclosure capture simulation data of gaze fixations while users are performing tasks using applications resident on the AR wearable device. The tilt angle of the AR display is selected based on including more gaze fixations that are within the field of view (FOV) of the AR display than are outside the FOV of the AR display. In some examples, an AR wearable device is manufactured with a fixed vertical tilt angle for the AR display. In some examples, the AR wearable device can dynamically adjust the vertical tilt angle of the AR display based on the applications that a user of the AR wearable device is likely to use or is using.
Systems and methods are provided for performing AR button selection operations on an augmented reality (AR) device. The system displays, by an AR device, a plurality of AR objects on a display region that overlaps a first real-world object, each of the plurality of AR objects being associated with an object selection region. The system computes a first spatial relationship factor for a first AR object of the plurality of AR objects based on a position of the first AR object relative to a position of a second real-world object and adjusts the object selection region of the first AR object based on the first spatial relationship factor. The system activates the first AR object in response to determining that the second real-world object overlaps the object selection region of the first AR object.
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
G06F 3/04815 - Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
G06T 7/70 - Determining position or orientation of objects or cameras
G06T 19/00 - Manipulating 3D models or images for computer graphics
G06T 19/20 - Editing of 3D images, e.g. changing shapes or colours, aligning objects or positioning parts
54.
MEDIA CONTENT ITEM GENERATION FOR A CONTENT SHARING PLATFORM
Systems and methods are provided for determining a set of selectors associated with the publisher identifier, each selector comprising specified content to extract from source data and one or more rules for extracting the specified content. The system and methods further provided for each location data in the list of location data, extracting, from the source data, specified content for each selector of at least a subset of the set of selectors based on the one or more rules specified in each selector of the at least the subset of the set of selectors; determining a template to use to generate the media content item, the template comprising regions corresponding to the one or more selectors; populating each region of the template using specified content for the corresponding selector; and generating the media content item from the populated template.
A system and a method for generating an automated GIF file generation system is described. In one aspect, the method includes accessing an animated GIF file, identifying a plurality of elements displayed in the animated GIF file, applying a variation of one or more elements to the animated GIF file, and generating a variant animated GIF file by applying the variation of the one or more elements to the animated GIF file. The system measures a trending metric of the variant animated GIF file based on a number of times the variant animated GIF file is shared on the communication platform and uses the trending metric as a feedback to generating the variant animated GIF file.
Aspects of the present disclosure involve a system comprising a computer-readable storage medium storing a program and method for presenting an overview of participant reactions to a virtual conference. The program and method provide for a virtual conference between plural participants; provide, for each of the plural participants, display of reaction buttons which are selectable by the participant to indicate different reactions to the virtual conference; receive indication of selections of the reaction buttons by one or more of the plural participants; store an indication of the selections over time in association with recording the virtual conference; generate a graphical overview of reactions to the virtual conference based on the stored indication of the selections; and provide, for a first participant of the plural participants, display of the graphical overview.
Provided are systems and methods for face reenactment. An example method includes receiving a target video that includes at least one target frame, where the at least one target frame includes a target face, receiving a scenario including a series of source facial expressions, determining, based on the target face, a target facial expression of the target face, synthesizing, based on a parametric face model and a texture model, an output face including the target face, where the target facial expression of the target face is modified to imitate a source facial expression of the series of source facial expressions, and generating, based on the output face, a frame of an output video. The parametric face model includes a template mesh pre-generated based on historical images of faces of a plurality of individuals, where the template mesh includes a pre-determined number of vertices.
A system and method for presentation of computer vision (e.g., augmented reality, virtual reality) using user data and a user code is disclosed. A client device can detect an image feature (e.g., scannable code) in one or more images. The image feature is determined to be linked to a user account. User data from the user account can then be used to generate one or more augmented reality display elements that can be anchored to the image feature in the one or more images.
G06T 11/60 - Editing figures and text; Combining figures or text
A63F 13/00 - Video games, i.e. games using an electronically generated display having two or more dimensions
A63F 13/213 - Input arrangements for video game devices characterised by their sensors, purposes or types comprising photodetecting means, e.g. cameras, photodiodes or infrared cells
A63F 13/352 - Interconnection arrangements between game servers and game devices; Interconnection arrangements between game devices; Interconnection arrangements between game servers - Details of game servers involving special game server arrangements, e.g. regional servers connected to a national server or a plurality of servers managing partitions of the game world
A63F 13/58 - Controlling game characters or game objects based on the game progress by computing conditions of game characters, e.g. stamina, strength, motivation or energy level
A63F 13/65 - Generating or modifying game content before or while executing the game program, e.g. authoring tools specially adapted for game development or game-integrated level editor automatically by game devices or servers from real world data, e.g. measurement in live racing competition
A63F 13/79 - Game security or game management aspects involving player-related data, e.g. identities, accounts, preferences or play histories
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
G06V 20/20 - Scenes; Scene-specific elements in augmented reality scenes
G06V 40/16 - Human faces, e.g. facial parts, sketches or expressions
H04L 51/52 - User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail for supporting social networking services
A contextual query system is configured to perform operations that include: causing display of a graphical user interface at a client device, the graphical user interface including a display of image data that comprises a set of image features; generating a query based on the set of image features of the image data; accessing media content based on the query at a repository, the repository comprising a collection of media content; and causing display of a presentation of the media content within the graphical user interface at the client device.
G06F 16/58 - Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
G06F 16/583 - Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
G06F 16/587 - Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using geographical or spatial information, e.g. location
60.
Normalized brightness control for user perception of visual media
A system and method for control of perceived brightness level in a display device based on the brightness of individual content items to provide a consistent viewing experience. This is a method of visual content-based brightness control. Adjusting the perceived display of individual visual content items does not rely on the adjustment of a display device's settings or controls, such as a general setting to control the brightness of a display screen, but rather evaluates the visual content being presented and adjusts the presentation layers for display of the visual content provided as part of the display screen. In this way, adjusting parameter(s) of one or more of these presentation layers provides control over the brightness level of the displayed content as a function of the brightness of the content.
Methods and systems for videoconferencing include generating work quality metrics based on emotion recognition of an individual such as a call center agent. The work quality metrics allow for workforce optimization. One example method includes the steps of receiving a video including a sequence of images, detecting an individual in one or more of the images, locating feature reference points of the individual, aligning a virtual face mesh to the individual in one or more of the images based at least in part on the feature reference points, dynamically determining over the sequence of images at least one deformation of the virtual face mesh, determining that the at least one deformation refers to at least one facial emotion selected from a plurality of reference facial emotions, and generating quality metrics including at least one work quality parameter associated with the individual based on the at least one facial emotion.
G06Q 10/0639 - Performance analysis of employees; Performance analysis of enterprise or organisation operations
G06V 40/16 - Human faces, e.g. facial parts, sketches or expressions
G10L 25/63 - Speech or voice analysis techniques not restricted to a single one of groups specially adapted for particular use for comparison or discrimination for estimating an emotional state
H04N 21/4402 - Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to MPEG-4 scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
G10L 15/22 - Procedures used during a speech recognition process, e.g. man-machine dialog
G10L 17/26 - Recognition of special voice characteristics, e.g. for use in lie detectors; Recognition of animal voices
H04N 21/442 - Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed or the storage space available from the internal hard disk
H04N 21/4788 - Supplemental services, e.g. displaying phone caller identification or shopping application communicating with other users, e.g. chatting
An Augmented Reality (AR) system provides stabilization of hand-tracking input data. The AR system provides for display a user interface of an AR application. The AR system captures, using one or more cameras of the AR system, video frame tracking data of a gesture being made by a user while the user interacts with the AR user interface. The AR system generates skeletal 3D model data of a hand of the user based on the video frame tracking data that includes one or more skeletal 3D model features corresponding to recognized visual landmarks of portions of the hand of the user. The AR system generates targeting data based on the skeletal 3D model data where the targeting data identifies a virtual 3D object of the AR user interface. The AR system filters the targeting data using a targeting filter component and provides the filtered targeting data to the AR application.
Aspects of the present disclosure involve a system comprising a computer-readable storage medium storing a program and method for presenting captured screen content within a virtual conferencing system. The program and method provide, in association with designing a room for virtual conferencing, a first interface for configuring at least one participant video element which is assignable to a respective participant video feed; receive, via the first interface, an indication of user input for setting first properties for the at least one participant video element; provide, in association with designing the room, a second interface for configuring presentation of screen content captured during virtual conferencing; receive, via the second interface, an indication of user input for setting second properties for the presentation of the screen content captured during virtual conferencing; and provide, in association with virtual conferencing, display of the room based on the first properties and the second properties.
H04L 12/18 - Arrangements for providing special services to substations for broadcast or conference
G06F 3/0484 - Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
64.
COORDINATING SIDE CONVERSATIONS WITHIN VIRTUAL CONFERENCING SYSTEM
Aspects of the present disclosure involve a system comprising a computer-readable storage medium storing a program and method for coordinating a side conversation within a virtual conferencing system. The program and method provide, for each of plural participants within a room, display of a participant video element correspond to the participant and including a user-selectable button to initiate a side conversation with the participant; receive indication of a user selection, by a first participant, of the user-selectable button for initiating the side conversation with a second participant; provide for a side conversation between a first device of the first participant and a second device of the second participant, the first and second participants maintaining presence in the room during the side conversation; and cause, for each of the plural participants, modified display of the participant video elements corresponding to the first and second participants, to indicate the side conversation.
H04L 12/18 - Arrangements for providing special services to substations for broadcast or conference
G06F 3/0484 - Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
65.
PRESENTING TIME-LIMITED VIDEO FEED WITHIN VIRTUAL WORKING ENVIRONMENT
Aspects of the present disclosure involve a system comprising a computer-readable storage medium storing a program and method for presenting a time-limited video feed within a virtual working environment. The program and method provide, within a virtual working environment for plural participants, display of an interface element which is user-selectable to capture video for sending to a group of participants within the virtual working environment, the group of participants having been selected from among the plural participants; receive, based on user selection of the interface element by a first participant in the group of participants, a time-limited video feed from a client device of the first participant; and provide, in response to the receiving and for each participant in the group of participants, display of the time-limited video feed.
Aspects of the present disclosure involve a system comprising a computer-readable storage medium storing a program and method for recreating keyboard and mouse sounds within a virtual working environment. The program and method provide for receiving, from a first client device of a first participant of a group of participants within a virtual working environment, a timing of keyboard and mouse input detected at the client device, the group of participants having been selected from among plural participants of the virtual working environment; generating, in response to the receiving, keyboard and mouse sounds that correspond to the timing of the keyboard and mouse input; and providing the generated keyboard and mouse sounds to one or more second client devices of respective one or more second participants of the group of participants, for presentation on the one or more second client devices.
An Augmented Reality (AR) system is provided. The AR system uses a combination of gesture and DMVO methodologies to provide for the user's selection and modification of virtual object of an AR experience. The user indicates that they want to interact with a virtual object of the AR experience by moving their hand to overlap the virtual object. While keeping their hand in an overlapping position, the user rotates their wrist and the virtual object is rotated as well. To end the interaction, the user moves their hand such that their hand is no longer overlapping the virtual object.
Systems, methods, and computer readable media for object counting on augmented reality (AR) wearable devices are disclosed. Embodiments are disclosed that enable display of a count of objects as part of a user view. Upon receipt of a request to count objects, the AR wearable device captures an image of the user view. The AR wearable device transmits the image to a backend for processing to determine the objects in the image. The AR wearable device selects a group of objects of the determined objects to count and overlays boundary boxes over counted objects within the user view. The position of the boundary boxes is adjusted to account for movement of the AR wearable device. A hierarchy of objects is used to group together objects that are related but have different labels or names.
Systems and methods are provided for performing operations on an augmented reality (AR) device using an external vision system. The system establishes, by the AR device, a communication with an external client device. The system overlays, by the AR device, a first AR object on a real-world environment being viewed using the AR device. The system receives interaction data from the external client device representing movement of a user determined by the external client device. The system, in response to receiving the interaction data from the external client device, modifies the first AR object by the AR device.
A system monitors a user environment via one or more sensors included in a computing device and detects, via a trigger, that event data is stored in a data store based on the monitoring. The system further detects one or more participants in the event data and invites the one or more participants to share an augmented reality event data and/or to a virtual reality event data. The system also creates, based on the event data, an augmented reality event data and/or a virtual reality event data, and presents the augmented reality event data and/or the virtual reality event data to the one or more participants in a synchronous mode and/or in an asynchronous mode, via the computing device.
Input indicative of a selection of volumetric content for presentation is received. The volumetric content comprises a volumetric representation of one or more elements of a real-world three-dimensional space. In response to the input, device state data associated with the volumetric content is accessed. The device state data describes a state of one or more network-connected devices associated with the real-world three-dimensional space. The volumetric content is presented. The presentation of the volumetric content includes presentation of the volumetric representation of the one or more elements overlaid on the real-world three-dimensional space by a display device and configuring the one or more network-connected devices using the device state data.
G06F 3/04815 - Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
The present disclosure relates to methods and systems for providing a multi-perspective augmented reality experience. A volumetric video of a three-dimensional space is captured. The volumetric video of the three-dimensional space includes a volumetric representation of a first user within the three-dimensional space. The volumetric video is displayed by a display device worn by a second user, and the second user sees the volumetric representation of the first user within the three-dimensional space. Input indicative of an interaction (e.g., entering or leaving) of the second user with the volumetric representation of the first user is detected. Based on detecting the input indicative of the interaction, the display device switches to a display of a recorded perspective of the first user. Thus, by interacting with a volumetric representation of the first user in a volumetric video, the second user views the first user's perspective of the three-dimensional space.
H04N 13/117 - Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation the virtual viewpoint locations being selected by the viewers or determined by viewer tracking
H04N 13/189 - Recording image signals; Reproducing recorded image signals
H04N 13/388 - Volumetric displays, i.e. systems where the image is built up from picture elements distributed through a volume
A system monitors an environment via one or more sensors included in a computing device and applies a trigger to detect that a memory experience is stored in a data store based on the monitoring. The system creates an augmented reality memory experience, a virtual reality memory experience, or a combination thereof, based on the trigger if the memory experience is detected. The system additionally projects the augmented reality memory experience, the virtual reality memory experience, or the combination thereof, via the computing device.
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
H04N 13/111 - Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation
H04N 13/239 - Image signal generators using stereoscopic image cameras using two 2D image sensors having a relative position equal to or related to the interocular distance
H04N 13/296 - Synchronisation thereof; Control thereof
H04N 13/388 - Volumetric displays, i.e. systems where the image is built up from picture elements distributed through a volume
The present disclosure relates to methods and systems for providing a touch-based augmented reality (AR) experience. During a capture phase, a first user may grip an object. An intensity of a force applied on the object in the grip and/or a duration of the grip may be recorded. A volumetric representation of the first user holding the object may also be captured. During an experience phase, a second user may touch the object, the object may provide haptic feedback (e.g., a vibration) to the second user at an intensity and a duration corresponding to an intensity of the force applied on the object and a duration of the grip of the object. If a volumetric representation of the first user holding the object is captured, touching the object may also cause a presentation of the first user's volumetric body that holds the object.
The present disclosure relates to methods and systems for providing a presentation of an experience (e.g., a journey) to a user using augmented reality (AR). During a capture phase, persons in the journey may take videos or pictures using their smartphones, GoPros, and/or smart glasses. A drone may also take videos or pictures during the journey. During an experience phase, an AR topographical rendering of the real-world environment of the journey may be rendered on a tabletop, highlighting/animating a path persons took in the journey. The persons may be rendered as miniature avatars/dolls overlaid on the representation of the real-world environment. When the user clicks on a point in the presentation of the journey, a perspective (e.g., the videos or pictures) at that point is presented.
A messaging system performs image processing to relight objects with neural networks for images provided by users of the messaging system. A method of relighting objects with neural networks includes receiving an input image with first lighting properties comprising an object with second lighting properties and processing the input image using a convolutional neural network to generate an output image with the first lighting properties and comprising the object with third lighting properties, where the convolutional neural network is trained to modify the second lighting properties to be consistent with lighting conditions indicated by the first lighting properties to generate the third lighting properties. The method further includes modifying the second lighting properties of the object to generate the object with modified second lighting properties and blending the third lighting properties with the modified second lighting properties to generate a modified output image comprising the object with fourth lighting properties.
Aspects of the present disclosure involve a system comprising a computer-readable storage medium storing a program and method for presenting participant reactions within a virtual working environment. The program and method provide a configuration interface for configuring a virtual working environment for plural participants, the configuration interface for specifying groups of participants, each group comprising respective participants selected from among the plural participants; receive first user input, provided via the configuration interface, specifying a first group of participants; provide, for each participant in the first group, display of a reactions interface with user-selectable buttons to indicate respective reactions for displaying to the first group; receive second user input, provided via the reactions interface, selecting one of the user-selectable buttons to indicate a reaction for displaying to the first group; and provide, for each participant in the first group, display of a reaction icon corresponding to the reaction.
G06F 3/04815 - Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
G06F 3/04817 - Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance using icons
The subject technology displays first augmented reality content on a computing device, the first augmented reality content comprising a first output media content. The subject technology provides for display a plurality of selectable graphical items, each of the selectable graphical items corresponding to a different augmented reality content including a set of media content modified utilizing facial synthesis. The subject technology receives a selection of one of the plurality of selectable graphical items. The subject technology, based at least in part on the selection, identifies second augmented reality content. The subject technology provides the second augmented reality content for display on the computing device.
A system and method for translating sign language utterances into a target language, including: receiving motion capture data; producing phonemes/sign fragments from the received motion capture data; producing a plurality of sign sequences from the phonemes/sign fragments; parsing these sign sequences to produce grammatically parsed sign utterances; translating the grammatically parsed sign utterances into grammatical representations in the target language; and generating output utterances in the target language based upon the grammatical representations.
A volumetric content presentation system includes a head-worn display device, which includes one or more processors, and a memory storing instructions that, when executed by the one or more processors, configure the display device to access AR content items that correspond to either real-world objects or virtual objects, mix and match these AR content items, and present volumetric content that includes these mixed and matched AR content items overlaid on a real-world environment to create a new AR scene that a user can experience.
G06T 19/00 - Manipulating 3D models or images for computer graphics
G06F 3/04815 - Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
81.
GENERATING IMMERSIVE AUGMENTED REALITY EXPERIENCES FROM EXISTING IMAGES AND VIDEOS
A two-dimensional element is identified from one or more two-dimensional images. A volumetric content item is generated based on the two-dimensional element identified from the one or more two-dimensional images. A display device presents the volumetric content item overlaid on a real-world environment that is within a field of view of a user of the display device.
G06V 10/26 - Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
G06V 10/94 - Hardware or software architectures specially adapted for image or video understanding
G06V 20/20 - Scenes; Scene-specific elements in augmented reality scenes
Systems, devices, media, and methods are presented for releasing an application feature in incremental stages while monitoring the application for anomalies. The feature includes a package of code and an action setting. The methods in some implementations include identifying active devices on which the application has been installed, monitoring the application according to a set of metrics, activating the feature by changing its action setting for a first segment of the active devices, pausing the feature if an anomaly is detected among the set of metrics, and generating a repair ticket. As long as no anomaly is detected, the activating step proceeds for subsequent segments of the active devices, iteratively, until the release is completed. A feature rank may be used to process and release a plurality of features in order of priority.
G06F 11/34 - Recording or statistical evaluation of computer activity, e.g. of down time, of input/output operation
G06F 11/14 - Error detection or correction of the data by redundancy in operation, e.g. by using different operation sequences leading to the same result
83.
COMBINING INDIVIDUAL FUNCTIONS INTO SHORTCUTS WITHIN A MESSAGING SYSTEM
Aspects of the present disclosure involve a system comprising a computer-readable storage medium storing a program and method for combining individual functions into shortcuts. The program and method provide for accessing information for a shortcut for performing with respect to an image captured by a device camera, the shortcut corresponding to a combination of individual functions to apply to the image; causing, based on the accessing, display of an interface with a user-selectable element for performing the shortcut with respect to the image; receiving, via the interface, user selection of the user-selectable element; and applying, in response to the receiving, the combination of the individual functions to the image.
In some implementations, a system may establish a video call between a first device associated with a first user and a second device associated with a second user of a communications platform. The system may present a video interface for the video call, the video interface comprising a first video stream generated by the first device of the first user and a second video stream generated by the second device associated with the second user. The system may present a first set of image augmentations selected by the communications platform in the video interface, the first set of video augmentations being selectable by the first user for augmentation of the first video stream generated by the first user device. The system may identify a second set of image augmentations used by a further set of users of the communications platform. The system may present the second set of image augmentations in the video interface, the second set of image augmentations being selectable by the first user for augmentation of the first video stream generated by the first user device.
Methods and systems are disclosed for performing operations for deforming an external mesh. The operations comprise receiving a video that includes a depiction of a real-world object. The operations comprise generating a three-dimensional (3D) body mesh associated with the real-world object that tracks movement of the real-world object across frames of the video. The operations comprise obtaining an external mesh associated with an Augmented-Reality (AR) element. The operations comprise accessing a plurality of deformation attributes associated with the external mesh, each attribute corresponding to a different deformation model. The operations comprise separately deforming, based on respective deformation models, a first portion of the external mesh and a second portion of the external mesh. The operations comprise modifying the video to include a display of the AR element based on the separately deformed first and second portions of the external mesh.
Aspects of the present disclosure involve a system comprising a computer-readable storage medium storing at least one program, method, and user interfaces to facilitate electronic chat conversations between users via multiple modes of electronic communication. A conversation view is displayed on a touch screen display of a client device. The conversation view presents messages exchanged between a first user and at least one other user in an electronic chat conversation. A user input comprising a touch gesture corresponding to an interaction with the display of the conversation view is received. Based on the user input, a camera feed view is displayed on the touch screen display. The camera feed view comprises a display of image data and enables a user to create a message comprising one or more images. A message is generated, and an indication of the message is displayed in the conversation view.
G06F 3/04845 - Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range for image manipulation, e.g. dragging, rotation, expansion or change of colour
G06F 3/04883 - Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
G06T 11/60 - Editing figures and text; Combining figures or text
H04M 1/7243 - User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
Aspects of the present disclosure involve a system and a method for performing operations comprising: receiving, by a messaging application, content from a given user; selecting a metric for measuring performance of the content on the messaging application; measuring performance of the content on the messaging application; computing a value of the performance of the content on the messaging application based on the selected metric; and updating a restricted use token wallet stored in a profile for the given user based on the computed value of the performance of the content.
H04L 51/046 - Interoperability with other network applications or services
A63F 13/87 - Communicating with other players during game play, e.g. by e-mail or chat
G06Q 20/36 - Payment architectures, schemes or protocols characterised by the use of specific devices using electronic wallets or electronic money safes
There is a remote control device or drone, which has software and a combination of lights or LED on an lighting ring or apparatus that can move independently of the drone; the drone can be programmed or be reactive to sound or other stimulus to create the effect of writing shapes or words in the air and typically at nighttime against a dark sky.
A wallpaper system presents a first wallpaper image of a wallpaper video to a user and receives, via a user input device, one or both of: (i) a spatial user input selection, and (ii) a time user input selection from the user to apply to the wallpaper video. In response to detecting one or both of: (i) the spatial user input selection, and (ii) the time user input selection, the wallpaper system determines one or both of: (i) a respective spatial movement parameter within a wallpaper video associated with the spatial user input selection, and (ii) a respective time coordinate within the wallpaper video associated with the time user input selection. Wallpaper system presents, via the image display, a second wallpaper image associated with one or both of: (i) the respective spatial movement parameter, and (ii) the respective time coordinate.
H04N 13/183 - On-screen display [OSD] information, e.g. subtitles or menus
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
G06F 3/04815 - Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
G06F 3/04883 - Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
An optical display system has an optical waveguide combiner and one or more cameras. The one or more camera(s) is optically coupled to the optical waveguide combiner and have a field of view of at least one real object and at least one virtual object displayable by the optical display system. The one or more camera(s), which may be for example wafer level waveguide camera(s), may be disposed outside the usable field of view of an output coupler. The one or more camera(s) may be self-calibrated electronically using images captured by the cameras of one or more virtual object(s) displayable by the optical display system. AR/VR/MR registration of devices and/or displayed virtual objects with real objects may be implemented using the images captured by the one or more camera(s) of the displayed virtual objects and real world objects. Real object distance and/or spatial location relative to the optical waveguide combiners may be determined or estimated from the captured images.
G06F 18/22 - Matching criteria, e.g. proximity measures
G06T 7/33 - Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
G06T 7/73 - Determining position or orientation of objects or cameras using feature-based methods
H04N 13/239 - Image signal generators using stereoscopic image cameras using two 2D image sensors having a relative position equal to or related to the interocular distance
Methods and systems are disclosed for performing operations comprising: accessing content received from a first client device associated with a first user; processing the content to identify a first image that depicts the first user wearing a first fashion item; determining a first pose of the first user depicted in the first image; searching a plurality of products to identify a first product that corresponds to the first pose of the first user depicted in the first image and that includes one or more attributes associated with the first fashion item; modifying the first image to generate an advertisement that depicts the first user wearing the first product; and during a content browsing session being accessed by the first client device, causing the advertisement that depicts the first user wearing the first product to be displayed automatically on the first client device.
Systems, methods, and computer readable media for voice-controlled user interfaces (UIs) for augmented reality (AR) wearable devices are disclosed. Embodiments are disclosed that enable a user to interact with the AR wearable device without using physical user interface devices. An application has a non-voice-controlled UI mode and a voice-controlled UI mode. The user selects the mode of the UI. The application running on the AR wearable device displays UI elements on a display of the AR wearable device. The UI elements have types. Predetermined actions are associated with each of the UI element types. The predetermined actions are displayed with other information and used by the user to invoke the corresponding UI element.
G06F 3/04817 - Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance using icons
G06F 3/0484 - Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
G10L 15/16 - Speech classification or search using artificial neural networks
G10L 15/22 - Procedures used during a speech recognition process, e.g. man-machine dialog
93.
COLLABORATIVE INTERFACE ELEMENT WITHIN A VIRTUAL CONFERENCING SYSTEM
Aspects of the present disclosure involve a system comprising a computer-readable storage medium storing a program and method for providing a collaborative interface element within a virtual conferencing system. The program and method provide, in association with designing a room, a first interface for configuring plural participant video elements, each being assignable to a respective participant; receive, via the first interface, an indication of user input for setting first properties for the plural participant video elements; provide a second interface for configuring a collaborative interface element within the room, and for specifying which of plural participant video elements permit navigating or modifying content within the collaborative interface element during virtual conferencing; receive, via the second interface, an indication of user input for setting second properties for the collaborative interface element; and provide, in association with virtual conferencing, display of the room based on the first properties and the second properties.
G06F 3/04847 - Interaction techniques to control parameter settings, e.g. interaction with sliders or dials
G06F 3/04815 - Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
H04L 65/401 - Support for services or applications wherein the services involve a main real-time session and one or more additional parallel real-time or time sensitive sessions, e.g. white board sharing or spawning of a subconference
H04L 65/403 - Arrangements for multi-party communication, e.g. for conferences
Aspects of the present disclosure involve a system comprising a computer-readable storage medium storing a program and method for virtual participant interaction for a hybrid event. The program and method provide an interface for designing a hybrid event between in-person and virtual participants, the interface for setting first properties of a physical space device present in a physical space, and setting second properties of a virtual room for the virtual participants; receive an indication of first user input setting the first and second properties; and in association with conferencing, configure the physical space device based on the first properties, display the virtual room based on the second properties, the virtual room including a live video feed of the physical space, receive, from a virtual participant device, an indication of second user input for interacting with the physical space device, and provide for interaction the based on the second user input.
Aspects of the present disclosure involve a system comprising a computer-readable storage medium storing a program and method for presenting change of presence sounds within a virtual working environment. The program and method provide for receiving, from a first client device of a first participant of a group of participants within a virtual working environment, interaction data relating to the first participant being present at the first client device, the group of participants having been selected from among plural participants of the virtual working environment; determining a change in presence of the first participant with respect to the first client device; generating, in response to the determining, a sound corresponding to the change in presence; and providing the generated sound to one or more second client devices of respective one or more second participants of the group of participants, for presentation on the one or more second client devices.
An Augmented Reality (AR) system is provided. The AR system uses a combination of gesture and DMVO methodologies to provide for the user's selection and modification of virtual objects of an AR experience. The user indicates that they want to interact with a virtual object of the AR experience by moving their hand to overlap the virtual object. While keeping their hand in an overlapping position, the user makes gestures that cause the user's viewpoint of the virtual object to either zoom in or zoom out. To end the interaction, the user moves their hand such that their hand is no longer overlapping the virtual object.
Systems and methods are provided for performing voice communication operations. The system establishes, by a first augmented reality (AR) device, a voice communication session between a plurality of users. The system displays, by the first AR device of a first user of the plurality of users, an avatar representing a second user of the plurality of users. The system receives, by the first AR device of a first user of the plurality of users, input from the first user that selects a display position for the avatar representing the second user within a real-world environment of the first user. The system animates the avatar representing the second user based on movement information received from a second AR device of the second user.
G06T 13/40 - 3D [Three Dimensional] animation of characters, e.g. humans, animals or virtual beings
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
G06F 3/0346 - Pointing devices displaced or positioned by the user; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
G06T 19/00 - Manipulating 3D models or images for computer graphics
G10L 25/78 - Detection of presence or absence of voice signals
H04L 65/1069 - Session establishment or de-establishment
98.
IN-PERSON PARTICIPANT INTERACTION FOR HYBRID EVENT
Aspects of the present disclosure involve a system comprising a computer-readable storage medium storing a program and method for in-person participant interaction for a hybrid event. The program and method provide an interface for designing a hybrid event between in-person and virtual participants, the interface for setting first properties of a physical space device present in a physical space, and setting second properties of a virtual room for the virtual participants; receive an indication of first user input setting the first and second properties; and in association with conferencing, configure the physical space device based on the first properties, display the virtual room based on the second properties, the virtual room including a live video feed of the physical space, receive, via a sidecar application accessible by the in-person participants, an indication of second user input for interacting with the physical space device, and provide for interaction based on the second user input.
A system captures via one or more sensors of a computing device, data of an environment observed by the one or more sensors at a first timeslot, and stores the data in a data store as a first portion of a timelapse memory experience. The system also captures, via the one or more sensors of a computing device, data of the environment observed by the one or more sensors at a second timeslot, and stores the data in a data store as a second portion of the timelapse memory experience. The system additionally associates the timelapse memory experience with a memory experience trigger, wherein the memory experience trigger can initiate a presentation of the timelapse memory experience.
Described herein is an application-to-friend data service for constructing and maintaining an application-to-friend index. The data service has a first service component that is triggered when an end-user opens an application. The first service component performs operations to update the database table with current data concerning other end-users who are friends of the end-user on whose behalf the data service has been invoked. The data service includes a second service component to receive and process a query by efficiently returning, for a given end-user and a given application, a list of connections or friends of the given end-user who are also end-users of the application.