A flexible electronic system comprises a flexible substrate comprising an interlayer elastomer dielectric and a plurality of rigid components disposed within the flexible substrate. The flexible electronic system may further comprise a high-density interconnect region comprising one or more routing layers. In examples, the flexible electronic system may further comprise a flexible barrier material encapsulating the flexible electronic system. In some examples the flexible electronic system and the high-density interconnect region may further comprise a plurality of routing layers and one or more through-vias. Each through-via may couple at least one of two rigid components, two routing layers, or a rigid component and a routing layer. In examples, at least some of the routing layers may comprise conductive traces.
An eye tracking system for a head-mounted devices includes an interference pattern emitter, an interference pattern detector, and processing logic. The interference pattern emitter provides at least two light beams that combine into a light pattern in an eyebox region of the head-mounted device. The light pattern includes an interference pattern based on constructive and destructive interference of the at least two light beams. The light detector is configured to detect a portion of the light pattern and provides detector data that is representative of one or more light intensities in the portion of the light pattern. The processing logic is coupled to the light detector to receive the detector data and is configured to identify displacement characteristics of the interference pattern. The processing logic is configured to determine orientation characteristics of an eye in the eyebox region based on the displacement characteristics.
In one embodiment, a method includes accessing a map of a building floor plan with locations of access points within the floor plan, the access points being capable of performing wireless communications with wireless devices. Determining a pose of a wireless device within the map using images captured by one or more cameras of the wireless device. Selecting a preferred access point based on the pose of the wireless device, the floor plan, and the locations of the plurality of access points within the floor plan. Configuring wireless communication settings of the wireless device to communicate with the preferred access point based on the pose of the wireless device and the location of the preferred access point within the floor plan.
The disclosed system may include at least one gradient-index liquid crystal lens. The system may include a selection module that selects a viewing angle. The system may also include an adjustment module that dynamically adjusts a phase reset property of the gradient-index liquid crystal lens in response to the selected viewing angle. Various other devices, systems, and methods are also disclosed.
G02F 1/29 - Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the position or the direction of light beams, i.e. deflection
5.
VIEW SYNTHESIS PIPELINE FOR RENDERING PASSTHROUGH IMAGES
A processor accesses a depth map and a first image of a scene generated using one or more sensors of an artificial reality device. The processor generates, based on the first image, segmentation masks respectively associated with a plurality of object types. The segmentation masks segment the depth map into a plurality of segmented depth maps respectively associated with the object types. The processor generates meshes using, respectively, the segmented depth maps. For each eye of the user, the processor captures a second image and generates, based on the second image, segmentation information. The processor warps the plurality of meshes to generate warped meshes for the eye, and then generates an eye-specific mesh for the eye by compositing the warped meshes according to the segmentation information. The processor renders an output image for the eye using the second image and the eye-specific mesh.
An actuator assembly includes a primary electrode, a secondary electrode overlapping at least a portion of the primary electrode, and an electroactive polymer layer disposed between the primary electrode and the secondary electrode, where the electroactive polymer layer includes a non-vertical (e.g., sloped) sidewall with respect to a major surface of at least one of the electrodes. The electroactive polymer layer may be characterized by a non-axisymmetric shape with respect to an axis that is oriented orthogonal to an electrode major surface.
H01L 41/317 - Applying piezo-electric or electrostrictive parts or bodies onto an electrical element or another base by depositing piezo-electric or electrostrictive layers, e.g. aerosol or screen printing by liquid phase deposition
H01L 41/09 - Piezo-electric or electrostrictive elements with electrical input and mechanical output
H01L 41/29 - Forming electrodes, leads or terminal arrangements
An eye tracking system employing three-dimensional (3D) sensing using time of flight is provided. Instead of directly measuring the time of arrival of the emitted photons, the light intensity of the transmitted laser beam is modulated, and a phase change of the return beam computed by comparison with the transmit waveform. The modulation frequency is resolved with radio frequency (RF) mixing and digital signal processing techniques. A variety of phase detection systems and techniques including, but not limited to, quadrature analog front end detection and analog homodyne phase detection are applied. The modulation frequency may, depending on phase detection technique, be sinusoidal or pulsed.
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
G01S 7/481 - Constructional features, e.g. arrangements of optical elements
G01S 7/4915 - Time delay measurement, e.g. operational details for pixel components; Phase measurement
G01S 17/32 - Systems determining position data of a target for measuring distance only using transmission of continuous waves, whether amplitude-, frequency-, or phase-modulated, or unmodulated
8.
Augment Orchestration in an Artificial Reality Environment
Aspects of the present disclosure are directed to an artificial reality system orchestrating interactions between virtual object “augments.” The orchestration can include linking, which can be forming two or more augments into a combination, embedding an augment within an existing combination, or triggering an action mapped to the linking of those augments. Another type of orchestration can include extracting, which can refer to taking an augment out of an existing combination, either by removing it from the combination or copying the augment to leave a version in the combination and having another version outside the combination.
G06F 3/04815 - Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
G06F 16/9536 - Search customisation based on social or collaborative filtering
G06T 19/00 - Manipulating 3D models or images for computer graphics
G06T 19/20 - Editing of 3D images, e.g. changing shapes or colours, aligning objects or positioning parts
A computer-implemented method, comprising accessing an image comprising a handheld device, wherein the image is captured by one or more cameras associated with the computing device, generating a cropped image that comprises a hand of a user or the handheld device from the image by processing the image, generating a vision-based six degrees of freedom (6DoF) pose estimation for the handheld device by processing the cropped image, metadata associated with the image, and first sensor data from one or more sensors associated with the handheld device, generating a map-based 6DoF pose estimation using the handheld device, and generating a final 6DoF pose estimation for the handheld device based on the vision-based 6DoF pose estimation and the map-based 6DoF pose estimation generated using the handheld device.
G06F 3/0346 - Pointing devices displaced or positioned by the user; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
G06F 3/03 - Arrangements for converting the position or the displacement of a member into a coded form
In one embodiment, a method by a computing system comprising a color camera and two monochrome cameras respectively associated with two eyes of a user includes computing a point cloud corresponding to a visible environment based at least on two stereoscopic grayscale images respectively captured by the two monochrome cameras, generating a mesh corresponding to the visible environment based on the computed point cloud, and generating two stereoscopic colorized images to be respectively displayed to the two eyes of the user, where each of the two stereoscopic colorized images is generated using (1) the mesh, (2) luminance information from one of the two stereoscopic grayscale images that is associated with the eye to which the stereoscopic colorized image is to be displayed, and (3) color information from a color input image captured by the color camera.
H04N 13/239 - Image signal generators using stereoscopic image cameras using two 2D image sensors having a relative position equal to or related to the interocular distance
G06T 7/593 - Depth or shape recovery from multiple images from stereo images
H04N 13/243 - Image signal generators using stereoscopic image cameras using three or more 2D image sensors
H04N 13/25 - Image signal generators using stereoscopic image cameras using image signals from one sensor to control the characteristics of another sensor
H04N 13/254 - Image signal generators using stereoscopic image cameras in combination with electromagnetic radiation sources for illuminating objects
H04N 13/344 - Displays for viewing with the aid of special glasses or head-mounted displays [HMD] with head-mounted left-right displays
A projector sequentially generates light pulses to form a sparse grid array. The light pulses reflect off objects in an environment and are reflected towards a sensor array. The sensor array sequentially senses the reflected light pulses across pixels of the sensor array. A depth sensing system calculates depth information of objects in the environment based on the sequential nature of the pulse generation and the pulse sensing. The depth sensing system calculates depth information based on both the positional and temporal information of generated light pulses, and the positional and temporal information of sensed light pulses. The depth sensing system may generate a representation of the environment based on the depth information.
G01B 11/25 - Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures by projecting a pattern, e.g. moiré fringes, on the object
G01S 17/89 - Lidar systems, specially adapted for specific applications for mapping or imaging
H04N 25/40 - Extracting pixel data from image sensors by controlling scanning circuits, e.g. by modifying the number of pixels sampled or to be sampled
H04N 25/60 - Noise processing, e.g. detecting, correcting, reducing or removing noise
12.
Electromagnetic interference reduction in extended reality environments
Methods and systems for reducing electromagnetic interference in analog circuit of a control device for an augmented reality (AR) or virtual reality (VR) system are described. An analog circuit associated with the control device may include at least one amplifier and an analog-to-digital converter coupled to an amplifier by one or more electrical conductors. Electromagnetic interference induced in the one or more electrical conductors by an external AC magnetic field may be reduced using at least one component of the control device configured to reduce the electromagnetic interference.
G09G 5/36 - Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators characterised by the display of individual graphic patterns using a bit-mapped memory
A61B 5/00 - Measuring for diagnostic purposes ; Identification of persons
A61B 5/296 - Bioelectric electrodes therefor specially adapted for particular uses for electromyography [EMG]
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
The disclosed computer-implemented method may include systems for generating personalized avatar reactions during live video broadcasts. For example, the systems and methods described herein can access a social networking system user's profile to identify an avatar associated with the social networking system user. The systems and methods can generate an avatar reaction by modifying one or more features of the avatar based on a corresponding emoticon reaction. Once generated, the social networking system user can select the avatar reaction for addition to an ephemeral reaction stream associated with a live video broadcast. Various other methods, systems, and computer-readable media are also disclosed.
In one embodiment, a method by an arbiter associated with hardware resources of a computing system includes associating with N indexed requesters requesting accesses to the hardware resources, where each of the N indexed requesters is associated with a credit counter and a weight, repeatedly granting a right to access the hardware resources to each requester that satisfies conditions in an indexing order among the N indexed requesters until none of the N indexed requesters satisfies the conditions and replenishing, upon a determination that none of the N indexed requesters satisfies the conditions, a credit counter associated with each of the N indexed requesters.
A thermoplastic interposer formed of ordered polymer sheets that may be configured to act as an interconnect and as a thermal energy spreader. In examples, the thermoplastic interposer may include one or more extensions or wings to further dissipate heat. In examples, laser direct structuring (LDS) may be used to form or more through silicon vias (TSVs) or through-chip vias. In examples, the ordered polymer sheets may be extruded via a roll-to-roll process, stacked, and processed to form a laminate interposer structure that makes up the interposer. In examples, the thermoplastic interposer may be used in multi-layer structures interconnecting two or more board assemblies such as printed circuit boards (PCB)s.
An electronic device includes a substrate and a circuit having a plurality of electrically-conductive components disposed on the substrate. The plurality of electrically-conductive components includes first, second and third electrically-conductive components. The third electrically-conductive component has a first end portion forming a first interface with the first electrically-conductive component and a second end portion forming a second interface with the second electrically conductive component. The first electrically-conductive component is made of a first material including a first metal. The second electrically-conductive component is made of a second material including the first metal. The third electrically-conductive component is made of a third material including a gallium-based alloy and a metallic filler. The metallic filler reduces a reactivity of the third electrically-conductive component with the first metal at the first and second interfaces, and thus minimizes deterioration of the first electrically-conductive component and the second electrically-conductive component over time.
H01L 23/29 - Encapsulation, e.g. encapsulating layers, coatings characterised by the material
H01B 1/22 - Conductive material dispersed in non-conductive organic material the conductive material comprising metals or alloys
H01L 21/48 - Manufacture or treatment of parts, e.g. containers, prior to assembly of the devices, using processes not provided for in a single one of the groups
A method and system for generating stylized representations in virtual/augmented reality applications. The method includes retrieving a two-dimensional (2D) image of a subject and a depth field associated with the 2D image. The method also includes generating a three-dimensional (3D) mesh based on the 2D image and the depth field. The method also includes generating a stylized texture field based on an analysis of the 2D image of the subject. The method also includes generating a 3D stylized model of the subject by enveloping the 3D mesh with the stylized texture field.
The disclosed method may include driving, using a biosensing device, a right-leg drive (RLD) signal, and measuring, using the biosensing device, a biosignal. The method also includes determining, from the measured biosignal, noise within the biosignal, and tuning, based on the noise, the RLD signal. Various other methods, systems, and computer-readable media are also disclosed.
G01D 3/036 - Measuring arrangements with provision for the special purposes referred to in the subgroups of this group mitigating undesired influences, e.g. temperature, pressure on measuring arrangements themselves
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
19.
View Synthesis Pipeline for Rendering Passthrough Images
A processor accesses a depth map and a first image of a scene generated using one or more sensors of an artificial reality device. The processor generates, based on the first image, segmentation masks respectively associated with a plurality of object types. The segmentation masks segment the depth map into a plurality of segmented depth maps respectively associated with the object types. The processor generates meshes using, respectively, the segmented depth maps. For each eye of the user, the processor captures a second image and generates, based on the second image, segmentation information. The processor warps the plurality of meshes to generate warped meshes for the eye, and then generates an eye-specific mesh for the eye by compositing the warped meshes according to the segmentation information. The processor renders an output image for the eye using the second image and the eye-specific mesh.
A method of forming a uniaxially oriented crystalline polymer article includes heating a segment of a crystallizable polymer article to a first temperature, applying a stress to the crystallizable polymer article in an amount effective to induce a positive strain within the segment of the crystallizable polymer article, and heating the segment of the crystallizable polymer article to a second temperature greater than the first temperature while continuing to apply the stress.
B29C 55/08 - Shaping by stretching, e.g. drawing through a die; Apparatus therefor of plates or sheets uniaxial, e.g. oblique transverse to the direction of feed
21.
SYSTEMS FOR CALIBRATING NEUROMUSCULAR SIGNALS SENSED BY A PLURALITY OF NEUROMUSCULAR-SIGNAL SENSORS, AND METHODS OF USE THEREOF
Methods for calibrating neuromuscular signals sensed by a plurality of neuromuscular-signal sensors are provided. One example method includes, in response to a triggering event, receiving a pulse sensed by a neuromuscular-signal sensor of a plurality of neuromuscular-signal sensors of a wrist-wearable device. The method also includes, in accordance with a determination that the pulse indicates that at least one neuromuscular-signal sensor of the plurality of neuromuscular-signal sensors of the wrist-wearable device is offset from a respective predetermined default position, determining a worn position of the wrist-wearable device on the user's wrist, and adjusting analysis of neuromuscular signals of the plurality of neuromuscular-signal sensors based on an offset between the worn position of the wrist-wearable device and a predetermined default position of the wrist-wearable device.
In one embodiment, a method by a computing system comprising a color camera and two monochrome cameras respectively associated with two eyes of a user includes computing a point cloud corresponding to a visible environment based at least on two stereoscopic grayscale images respectively captured by the two monochrome cameras, generating a mesh corresponding to the visible environment based on the computed point cloud, and generating two stereoscopic colorized images to be respectively displayed to the two eyes of the user, where each of the two stereoscopic colorized images is generated using (1) the mesh, (2) luminance information from one of the two stereoscopic grayscale images that is associated with the eye to which the stereoscopic colorized image is to be displayed, and (3) color information from a color input image captured by the color camera.
A distributed imaging system for augmented reality devices is disclosed. The system includes a computing module in communication with a plurality of spatially distributed sensing devices. The computing module is configured to process input images from the sensing devices based on performing a local feature matching computation to generate corresponding first output images. The computing module is further configured to process the input images based on performing an optical flow correspondence computation to generate corresponding second output images. The computing module is further configured to computationally combine first and second output images to generate third output images.
A projector sequentially generates light pulses to form a sparse grid array. The light pulses reflect off objects in an environment and are reflected towards a sensor array. The sensor array sequentially senses the reflected light pulses across pixels of the sensor array. A depth sensing system calculates depth information of objects in the environment based on the sequential nature of the pulse generation and the pulse sensing. The depth sensing system calculates depth information based on both the positional and temporal information of generated light pulses, and the positional and temporal information of sensed light pulses. The depth sensing system may generate a representation of the environment based on the depth information.
A tunable lens including a pair of reflectors is disclosed. At least one of the reflectors may be curved for contributing to the focusing or defocusing power of the lens. At least one of the reflectors is translatable for tuning the focusing/defocusing power. The reflectors may be configured in a pancake lens configuration where one of the reflectors is a 50/50 reflector and the other is a polarization selective reflector. Refractive elements may be disposed between the reflectors for providing more optical power to the lens, and/or for balancing optical aberrations.
G02B 26/08 - Optical devices or arrangements for the control of light using movable or deformable optical elements for controlling the direction of light
A graded-index optical element may include a nanovoided material including a first surface and a second surface opposite the first surface. The nanovoided material may be transparent between the first surface and the second surface. Additionally, the nanovoided material may have a predefined change in effective refractive index in at least one axis due to a change in at least one of nanovoid size or nanovoid distribution along the at least one axis. Various other elements, devices, systems, materials, and methods are also disclosed.
B29D 11/00 - Producing optical elements, e.g. lenses or prisms
G02B 1/04 - Optical elements characterised by the material of which they are made; Optical coatings for optical elements made of organic materials, e.g. plastics
Various embodiments set forth high-resolution liquid crystal displays and components thereof. In some embodiments, light emitted by a high-resolution green color liquid crystal display is combined, via a combiner, with light emitted by at least one lower-resolution red and blue color liquid crystal display. The red and blue color display(s) may include a single display or two displays positioned on opposing sides of the combiner. The combiner may be a dichroic or polarization-based combiner. Combined light from the green color display and the red and blue color display(s) is passed through collimating optics, such as a pancake lens or a Fresnel lens, toward a viewer's eye.
Methods and corresponding systems and apparatuses for saving power through selectively disabling clock signals in a systolic array are described. In some embodiments, a clock gate controller is operable to output a gated clock signal from which local clock signals of processing elements in the systolic array are derived. The gated clock signal corresponds to a root clock signal that is distributed through a clock distribution network or clock tree. The clock gate controller is located along one branch of the clock distribution network. The branch can be associated with processing elements that form a column within the systolic array. Disabling the gated clock signal disables the local clock signals along the entire branch, preventing any components that are clocked by those local clock signals from consuming power. Additional clock gate controllers can similarly be provided for other branches, including a branch associated with another column.
G06F 1/00 - ELECTRIC DIGITAL DATA PROCESSING - Details not covered by groups and
G06F 1/3237 - Power saving characterised by the action undertaken by disabling clock generation or distribution
G06F 15/80 - Architectures of general purpose stored program computers comprising an array of processing units with common control, e.g. single instruction multiple data processors
G06F 1/3203 - Power management, i.e. event-based initiation of a power-saving mode
29.
ADAPTIVE IMPEDANCE TUNING FOR CAPACITIVE INTERCONNECTS
The disclosed system may include a detachable capsule that includes a capsule-side capacitive plate. The system may also include a receiving antenna electrically connected to an antenna-side capacitive plate, and a capacitive sensor electrically connected to the capsule-side capacitive plate. The two plates may be coupled to each other and may transmit RF signals between them. The capacitive sensor may be configured to detect an amount of capacitance between the two capacitive plates. The system may also include an antenna matching tuner electrically connected to the capacitive sensor and to an antenna feed. Then, upon receiving capacitance measurements from the capacitive sensor, the antenna matching tuner may alter various parameters of the antenna feed including impedance matching parameters. Various other apparatuses and mobile wearable devices are also disclosed.
H01Q 1/27 - Adaptation for use in or on movable bodies
G01R 27/26 - Measuring inductance or capacitance; Measuring quality factor, e.g. by using the resonance method; Measuring loss factor; Measuring dielectric constants
G04G 21/04 - Input or output devices integrated in time-pieces using radio waves
H01Q 7/00 - Loop antennas with a substantially uniform current distribution around the loop and having a directional radiation pattern in a plane perpendicular to the plane of the loop
Aspects of the present disclosure are directed to creating virtual doors within artificial reality (XR) universes for traversal within that XR universe and between other XR universes. Users can create virtual doors that control access to their privately owned property (e.g., world, parcel, house, etc.) in an XR universe. For example, an owner of a virtual door can manually lock the door to prevent any user from entering their property through the virtual door. As another example, an owner of a virtual door can configure door permissions and/or privacy settings that serve as heuristics by which a door access control manager determines whether to authorize a particular user, XR world, and/or XR universe to access. The XR universe traversal system can control an execution environment to smoothly transition between different applications, thereby enabling a user to traverse between different XR universes without having to leave the XR environment.
A room manager can generate mappings for a real-world room that support a shared XR environment. For example, the real-world room can include real-world objects and surfaces, such as a table(s), chair(s), wall(s), door(s), window(s), etc. The room manager can generate XR object definitions based on information received about the real-world room, object(s), and surface(s). For example, the room manager can implement a flow that guides a user equipped with an XR system to provide information for the XR object definitions, such as real-world surfaces that map to the XR object(s), borders (e.g., measured using a component of the XR system), such as borders on real-world surfaces, semantic information (e.g., number of seat assignments at an XR table, size of XR objects, etc.), and other suitable information. Implementations generate previews of the shared XR environment, such as a local preview and a remote preview.
The disclosed system may include a conductive enclosure, a first printed circuit board (PCB) that includes multiple antenna feeds, and a second PCB that includes a grounding layer and one or more sensors. A first antenna feed may be electrically connected to the conductive enclosure, and a second antenna feed may be electrically connected to the grounding layer of the second PCB. As such, the grounding layer of the second PCB may act as a radiating element for a second antenna. Various other mobile electronic devices, apparatuses, and methods of manufacturing are also disclosed.
H01Q 5/328 - Individual or coupled radiating elements, each element being fed in an unspecified way using frequency dependent circuits or components, e.g. trap circuits or capacitors between a radiating element and ground
H01Q 1/52 - Means for reducing coupling between antennas; Means for reducing coupling between an antenna and another structure
A camera assembly with two distinct magnifications is described. The camera assembly includes a lens assembly with two modules to perceive light. The first module includes two halves of different lens assemblies disposed together to provide two different optical paths and fields of view to obtain the two different optical magnifications. The second module includes a lens assembly common to the two halves of the first module. A biprism provides light separation for the two halves of the first module. Thus, light received and subjected to two different magnifications in the first module at a distal end of the camera assembly is further processed through the second module before being provided to a sensor. The biprism at the camera's distal end keeps the two different lens assemblies of the first module at the same viewing direction.
H04N 23/55 - Optical parts specially adapted for electronic image sensors; Mounting thereof
G02B 13/18 - Optical objectives specially designed for the purposes specified below with lenses having one or more non-spherical faces, e.g. for reducing geometrical aberration
G02B 27/00 - Optical systems or apparatus not provided for by any of the groups ,
G06F 1/16 - Constructional details or arrangements
H04N 5/262 - Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects
34.
SYSTEMS, METHODS, AND DEVICES FOR PRODUCING INTERCONNECTS ON DEFORMABLE SUBSTRATES OF ELECTRONIC DEVICES
The present disclosure provides systems, methods, and devices for producing an interconnect. An electronic device of the present disclosure includes a deformable substrate including a circuit. The circuit includes a channel extending from a first portion of the deformable substrate to a second portion of the deformable substrate. A first circuit component is adjacent to the first portion of the deformable substrate. A second circuit component is adjacent to the second portion of the deformable substrate. A first metal material is formed overlaying a first portion of the deformable substrate including a first portion of the channel. A second metal material interfaces with the first metal material, thereby substantially occupying an interior volume of the channel.
H05K 3/32 - Assembling printed circuits with electric components, e.g. with resistor electrically connecting electric components or wires to printed circuits
H05K 3/18 - Apparatus or processes for manufacturing printed circuits in which conductive material is applied to the insulating support in such a manner as to form the desired conductive pattern using precipitation techniques to apply the conductive material
Aspects of the present disclosure are directed to creating virtual doors within artificial reality (XR) universes for traversal within that XR universe and between other XR universes. Users can create virtual doors that control access to their privately owned property (e.g., world, parcel, house, etc.) in an XR universe. For example, an owner of a virtual door can manually lock the door to prevent any user from entering their property through the virtual door. As another example, an owner of a virtual door can configure door permissions and/or privacy settings that serve as heuristics by which a door access control manager determines whether to authorize a particular user, XR world, and/or XR universe to access. The XR universe traversal system can control an execution environment to smoothly transition between different applications, thereby enabling a user to traverse between different XR universes without having to leave the XR environment.
A method and system for text-to-video generation. The method includes receiving a text input, generating a representation frame based on the text input using a model trained on text-image pairs, generating a set of frames based on the representation frame and a first frame rate, interpolating the set of frames to a higher frame rate, generating a first video based on the interpolated set of frames, increasing a resolution of the first video based on a first and second super-resolution model, and generating an output video based on a result of the super-resolution models.
Methods, systems, and storage media for generating audio data includes receiving a text input. The method also includes receiving a plurality of representative audio sources and encoding the plurality of representative audio sources into a plurality of audio tokens. The method includes encoding the text input into a plurality of text representations. The method comprises mapping each audio tokens of the plurality of audio tokens to a text representation of the plurality of text representations. The method also comprises determining a relationship score based on mapping each audio tokens to the text representation, wherein the relationship score identifies a distribution of audio tokens from the plurality of audio tokens. The method and systems can also comprise decoding the subgroup of audio tokens to yield a reconstmcted audio source.
G10L 13/08 - Text analysis or generation of parameters for speech synthesis out of text, e.g. grapheme to phoneme translation, prosody generation or stress or intonation determination
In one example, an apparatus for integrating sensing and display system includes a first semiconductor layer that includes an image sensor; a second semiconductor layer that includes a display; a third semiconductor layer that includes compute circuits configured to support an image sensing operation by the image sensor and a display operation by the display; and a semiconductor package that encloses the first, second, and third semiconductor layers, the semiconductor package further including a first opening to expose the image sensor and a second opening to expose the display. The first, second, and third semiconductor layers form a first stack structure along a first axis. The third semiconductor layer is sandwiched between the first semiconductor layer and the second semiconductor layer in the first stack structure.
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
G06V 10/25 - Determination of region of interest [ROI] or a volume of interest [VOI]
G09G 3/32 - Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes for presentation of an assembly of a number of characters, e.g. a page, by composing the assembly by combination of individual elements arranged in a matrix using controlled light sources using electroluminescent panels semiconductive, e.g. using light-emitting diodes [LED]
H01L 23/00 - SEMICONDUCTOR DEVICES NOT COVERED BY CLASS - Details of semiconductor or other solid state devices
H01L 27/15 - Devices consisting of a plurality of semiconductor or other solid-state components formed in or on a common substrate including semiconductor components with at least one potential-jump barrier or surface barrier, specially adapted for light emission
H04N 5/232 - Devices for controlling television cameras, e.g. remote control
Aspects of the present disclosure are directed to systems for capturing, recording, and playing back steps for building objects and structures in an artificial reality (XR) world. A teacher or creator can initiate a capture context within a build mode of an XR application, during which steps (and potentially other metadata) of a building process are recorded and stored. A XR world building engine can generate an interactive replay of the recorded building process that allows another user to view the building process within an XR environment at the user's preferred pace and/or from different perspectives. In some implementations, the XR world building engine can generate a replay of the teacher's or creator's avatar performing the building process, such that the user can view the building process from a third person perspective and perform the same steps to learn how to build objects and structures in an XR world.
The disclosed device may include a lens stack. The lens stack may include a first gradient-index liquid crystal lens and a second gradient-index liquid crystal lens in tandem with the first gradient-index liquid crystal lens. The lens stack may be configured to reach a target optical power based on a first optical power of the first gradient-index liquid crystal lens and a second optical power of the second gradient-index liquid crystal lens. Various other devices, systems, and methods are also disclosed.
G02F 1/29 - Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the position or the direction of light beams, i.e. deflection
A device is provided. The device includes a waveguide configured to guide an image light to propagate from a light inputting surface to a light outputting surface. The waveguide includes a substrate having a back surface facing an eye-box region of the device and a front surface opposite to the back surface, a plurality of out-coupling structures disposed at the back surface or at least partially inside the substrate, and a medium layer embedded inside the substrate between the out-coupling structures and the front surface. The medium layer has a refractive index that is lower than the substrate. The device also includes an optical lens printed over the back surface of the substrate.
Methods, systems, and storage media for generating audio data includes receiving a text input. The method also includes receiving a plurality of representative audio sources and encoding the plurality of representative audio sources into a plurality of audio tokens. The method includes encoding the text input into a plurality of text representations. The method comprises mapping each audio tokens of the plurality of audio tokens to a text representation of the plurality of text representations. The method also comprises determining a relationship score based on mapping each audio tokens to the text representation, wherein the relationship score identifies a distribution of audio tokens from the plurality of audio tokens. The method and systems can also comprise decoding the subgroup of audio tokens to yield a reconstructed audio source.
G10L 19/018 - Audio watermarking, i.e. embedding inaudible data in the audio signal
G10L 19/02 - Speech or audio signal analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using spectral analysis, e.g. transform vocoders or subband vocoders
An image-sensing system includes multiple sensing modules. Each of the multiple sensing modules includes multiple optical sensors arranged in an array. Each of the multiple sensors is configured to be switched on and off to generate analog sensing data. The image-sensing system also includes an analog-to-digital converter (ADC) shared by the multiple optical sensors configured to convert analog sensing data generated by the multiple optical sensors into digital data. The image-sensing system also includes a processor configured to control the multiple sensing modules.
An administered authentication system can authenticate an artificial reality device using an authorization record between a user account and an artificial reality device. In some implementations, the authorization record is created in response to activation of a user account-specific key sent to a user-supplied contact, where an artificial reality device identifier was provided with the user-supplied contact. In other implementations, the authorization record is created in response to activation of a user account-specific key provided to the artificial reality device as a code, where activation of the key includes adding an artificial reality device identifier to a key activation message. In yet other implementations, the authorization record is created in response to an application associated with a user account activating an artificial reality device-specific key, with an artificial reality device identifier, that is provided via the artificial reality device.
G06F 21/34 - User authentication involving the use of external additional devices, e.g. dongles or smart cards
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
G06F 21/33 - User authentication using certificates
G06F 21/36 - User authentication by graphic or iconic representation
G06F 21/40 - User authentication by quorum, i.e. whereby two or more security principals are required
G06T 19/00 - Manipulating 3D models or images for computer graphics
H04L 9/32 - Arrangements for secret or secure communications; Network security protocols including means for verifying the identity or authority of a user of the system
45.
Translating Interactions on a Two-Dimensional Interface to an Artificial Reality Experience
Many users access artificial reality (XR) experiences through their mobile phones. However, it is difficult to translate XR experiences to a two-dimensional (2D) screen in a way that feels intuitive and natural. Thus, the technology can map an interaction plane in a three-dimensional (3D) scene to the 2D screen with as many affordances as possible, even if the plane is not parallel to the mobile phone. The plane can be a fixed surface or a dynamically changeable surface through which the user sends inputs through the 2D screen. The mapping of the plane to the 2D screen can control interaction with a virtual object on the interaction plane in the XR environment, enabling parity between the same experience on XR and non-XR interfaces.
G06F 3/04815 - Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
G06F 3/0346 - Pointing devices displaced or positioned by the user; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
G06F 3/04842 - Selection of displayed objects or displayed text elements
G06F 3/04845 - Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range for image manipulation, e.g. dragging, rotation, expansion or change of colour
G06F 3/04886 - Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures by partitioning the display area of the touch-screen or the surface of the digitising tablet into independently controllable areas, e.g. virtual keyboards or menus
The disclosed computer-implemented method may include coupling an array of reflective polarizers to a back surface of a back optical substrate; coupling an array of quarter-wave plates to a front surface of the back optical substrate such that the array of quarter-wave plates is aligned with the array of reflective polarizers; molding the back optical substrate with at least one initial mold, wherein the at least one initial mold defines an initial array of optical element surfaces that is aligned with the array of quarter-wave plates; and twin-sheet thermoforming, between a front twin-sheet mold and a back twin-sheet mold, the back optical substrate with a front optical substrate, wherein the front twin-sheet mold defines a front array of optical element surfaces and the back mold defines a back array of optical element surfaces. Various other methods, apparatuses, and systems are also disclosed.
A room manager can generate mappings for a real-world room that support a shared XR environment. For example, the real-world room can include real-world objects and surfaces, such as a table(s), chair(s), wall(s), door(s), window(s), etc. The room manager can generate XR object definitions based on information received about the real-world room, object(s), and surface(s). For example, the room manager can implement a flow that guides a user equipped with an XR system to provide information for the XR object definitions, such as real-world surfaces that map to the XR object(s), borders (e.g., measured using a component of the XR system), such as borders on real-world surfaces, semantic information (e.g., number of seat assignments at an XR table, size of XR objects, etc.), and other suitable information. Implementations generate previews of the shared XR environment, such as a local preview and a remote preview.
A room manager can generate mappings for a real-world room that support a shared XR environment. For example, the real-world room can include real-world objects and surfaces, such as a table(s), chair(s), wall(s), door(s), window(s), etc. The room manager can generate XR object definitions based on information received about the real-world room, object(s), and surface(s). For example, the room manager can implement a flow that guides a user equipped with an XR system to provide information for the XR object definitions, such as real-world surfaces that map to the XR object(s), borders (e.g., measured using a component of the XR system), such as borders on real-world surfaces, semantic information (e.g., number of seat assignments at an XR table, size of XR objects, etc.), and other suitable information. Implementations generate previews of the shared XR environment, such as a local preview and a remote preview.
In one embodiment, a method includes receiving, by a first client system, from one or more remote servers, a current version of a neural network model including multiple model parameters, training the neural network model on multiple examples retrieved from a local data store to generate multiple updated model parameters, wherein each of the examples includes one or more features and one or more labels, calculating a user valuation associated with the first client system, wherein the user valuation represents a measure of utility of training the neural network model on the multiple examples, and sending, to one or more of the remote servers, the trained neural network model and the user valuation, wherein the user valuation is associated with a likelihood of the first client system being selected for a subsequent training of the neural network model.
A room manager can generate mappings for a real-world room that support a shared XR environment. For example, the real-world room can include real-world objects and surfaces, such as a table(s), chair(s), wall(s), door(s), window(s), etc. The room manager can generate XR object definitions based on information received about the real-world room, object(s), and surface(s). For example, the room manager can implement a flow that guides a user equipped with an XR system to provide information for the XR object definitions, such as real-world surfaces that map to the XR object(s), borders (e.g., measured using a component of the XR system), such as borders on real-world surfaces, semantic information (e.g., number of seat assignments at an XR table, size of XR objects, etc.), and other suitable information. Implementations generate previews of the shared XR environment, such as a local preview and a remote preview.
Many users access artificial reality (XR) experiences through their mobile phones. However, it is difficult to translate XR experiences to a two-dimensional (2D) screen in a way that feels intuitive and natural. Thus, the technology can map an interaction plane in a three-dimensional (3D) scene to the 2D screen with as many affordances as possible, even if the plane is not parallel to the mobile phone. The plane can be a fixed surface or a dynamically changeable surface through which the user sends inputs through the 2D screen. The mapping of the plane to the 2D screen can control interaction with a virtual object on the interaction plane in the XR environment, enabling parity between the same experience on XR and non-XR interfaces.
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
G06F 3/0346 - Pointing devices displaced or positioned by the user; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
G06F 3/04815 - Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
Aspects of the present disclosure are directed to systems for capturing, recording, and playing back steps for building objects and structures in an artificial reality (XR) world. A teacher or creator can initiate a capture context within a build mode of an XR application, during which steps (and potentially other metadata) of a building process are recorded and stored. A XR world building engine can generate an interactive replay of the recorded building process that allows another user to view the building process within an XR environment at the user's preferred pace and/or from different perspectives. In some implementations, the XR world building engine can generate a replay of the teacher's or creator's avatar performing the building process, such that the user can view the building process from a third person perspective and perform the same steps to learn how to build objects and structures in an XR world.
Aspects of the present disclosure are directed to streaming interactive content from a native application executing at an artificial reality (XR) device into an artificial reality environment and/or to nearby XR device(s). A shell environment at an XR system can manage the software components of the system. The shell environment can include a shell application and a three-dimensional shell XR environment displayed to a user. An additional application, natively executing at the XR system, can provide a host version of content and a remote version of content. A two-dimensional virtual object displayed in the shell XR environment can display the host version of the content, and the remote version of the content can be streamed to a remote XR system. The remote XR system can display the remote content within another two-dimensional virtual object, for example in another shell XR environment displayed by the remote XR system.
A system on a chip includes a first subsystem comprising a first memory; a second subsystem comprising a second memory; and an always-on subsystem. The always-on subsystem can comprise processing circuitry configured to: in response to a first activation event, signal the first subsystem to initiate repair operations on the first memory, and in response to a second activation event occurring after the first event, signal the second subsystem to initiate repair operations on the second memory.
In some implementations, the disclosed systems and methods can detect an interaction with respect to a set of virtual objects, which can start with a particular gesture, and take an action with respect to one or more virtual objects based on a further interaction (e.g., holding the gesture for a particular amount of time, moving the gesture in a particular direction, releasing the gesture, etc.).
In some implementations, the disclosed systems and methods can detect an interaction with respect to a set of virtual objects, which can start with a particular gesture, and take an action with respect to one or more virtual objects based on a further interaction (e.g., holding the gesture for a particular amount of time, moving the gesture in a particular direction, releasing the gesture, etc.).
In some implementations, the disclosed systems and methods can automatically review a 3D video to determine a depicted user or avatar movement pattern (e.g., dance moves, repair procedure, playing an instrument, etc.).
In some implementations, the disclosed systems and methods can detect an interaction with respect to a set of virtual objects, which can start with a particular gesture, and take an action with respect to one or more virtual objects based on a further interaction (e.g., holding the gesture for a particular amount of time, moving the gesture in a particular direction, releasing the gesture, etc.).
In some implementations, the disclosed systems and methods can automatically review a 3D video to determine a depicted user or avatar movement pattern (e.g., dance moves, repair procedure, playing an instrument, etc.).
In some implementations, the disclosed systems and methods can allow the gesture to included a flat hand with the user's thumb next to the palm, with the gesture toward the user's face.
In particular embodiments, a computing system may access a particular image frame and corresponding depth information of a dynamic scene. The depth information is used to generate a point cloud of the particular image frame. The system may generate a first latent representation based on the point cloud. The system may access a sequence of image frames of the dynamic scene and a set of key frames. The system may generate, using a temporal transformer, a second latent representation based on tracking and combining temporal relationship between the sequence of image frames and the set of key frames. The system may access camera parameters for rendering the one or more objects from a desired novel viewpoint and generate a third latent representation. The system may train an improved neural radiance fields (NeRF) based model for free-viewpoint rendering of the dynamic scene based on the first, second, and third latent representations.
A display engine includes an arrayed light source panel including a light source array of individually addressable light sources, each light source being configured to emit a first light beam associated with a first wavelength band. The display engine also includes a beam reshaping module including beam reshaping elements, each beam reshaping element being configured to reshape a first beam profile of the first light beam and output a second light beam with a second beam profile. The display engine also includes a transmissive display driver panel including a display driver module integrated with a pixelated color conversion module that includes a plurality of color conversion units configured to at least partially convert the second light beam into a third light beam associated with a second wavelength band. The display engine also includes an active light modulation medium configured to modulate the third light beam for displaying an image.
The disclosed apparatus may include a wearable haptic ring that features input capabilities relative to a computing system. In various examples, the wearable haptic ring may be designed to curve around a human finger of a wearer with a touchpad that is seamlessly integrated with the ring. For example, the seamlessly integrated touchpad may be operable by another finger of the wearer. Moreover, the haptic ring may include a haptic feedback unit designed to provide haptic feedback in response to input from the wearer. As such, the haptic ring may enable a wide range of user inputs while appearing like a typical ring rather than a computer input/output device. Various other implementations are also disclosed.
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
G06F 3/0354 - Pointing devices displaced or positioned by the user; Accessories therefor with detection of 2D relative movements between the device, or an operating part thereof, and a plane or surface, e.g. 2D mice, trackballs, pens or pucks
59.
HYBRID ELECTRICALLY AND THERMALLY SWITCHABLE SYSTEM USING HEAT SOURCE
An optical device includes a first optically dimmable switch for providing a first transmittance while the first optically dimmable switch is in a first state and providing a second transmittance distinct from the first transmittance while the first optically dimmable switch is in a second state distinct from the first state. The optical device also includes a dynamic heat source thermally coupled with the first optically dimmable switch. The dynamic heat source is at a first temperature at a first time and is at a second temperature distinct from the first temperature at a second time mutually exclusive from the first time. The optical device may operate as an optical dimming device, which may be used in head-mounted display devices or as dimmable windows or shutters.
G02F 1/13 - Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the intensity, phase, polarisation or colour based on liquid crystals, e.g. single liquid crystal display cells
H01H 61/02 - Electrothermal relays wherein the thermally-sensitive member is heated indirectly, e.g. resistively, inductively
60.
User Controlled Task Execution with Task Persistence for Assistant Systems
In one embodiment, a method includes receiving a first user request at a client system to suspend a first task being executed by an assistant system operating on the client system, suspending the execution of the first task responsive to the first user request, receiving a second user request at the client system, determining that the second user request is a request to resume the suspended first task based on user interactions with the assistant system with respect to one or more entities associated with the first task, and presenting a prompt to resume the first task at the client system.
G06V 10/764 - Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
G06V 10/82 - Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
H04L 51/212 - Monitoring or handling of messages using filtering or selective blocking
H04L 51/222 - Monitoring or handling of messages using geographical location information, e.g. messages transmitted or received in proximity of a certain spot or area
H04L 51/224 - Monitoring or handling of messages providing notification on incoming messages, e.g. pushed notifications of received messages
H04L 51/52 - User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail for supporting social networking services
The disclosed computer-implemented method may include applying, via a sound reproduction system, sound cancellation that reduces an amplitude of various sound signals. The method further includes identifying, among the sound signals, an external sound whose amplitude is to be reduced by the sound cancellation. The method then includes analyzing the identified external sound to determine whether the identified external sound is to be made audible to a user and, upon determining that the external sound is to be made audible to the user, the method includes modifying the sound cancellation so that the identified external sound is made audible to the user. Various other methods, systems, and computer-readable media are also disclosed.
G10K 11/178 - Methods or devices for protecting against, or for damping, noise or other acoustic waves in general using interference effects; Masking sound by electro-acoustically regenerating the original acoustic waves in anti-phase
H04R 1/40 - Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by combining a number of identical transducers
An example method for execution on a system on a chip (SoC) having a plurality of subsystems includes receiving, by a storage controller from a subsystem of the plurality of subsystems, a command to fetch, from a local memory, task descriptor data comprising access parameters for accessing a storage device, the access parameters including a storage device address; obtaining, by an encryption engine of the SoC, the command to fetch the task descriptor data; determining, by the encryption engine based on an access rule, whether the subsystem has sufficient privilege to access the storage device address; in response to determining that the subsystem has sufficient privilege to access the storage device, encrypting, source data in the local memory according to an encryption key associated with the subsystem; and providing the encrypted source data to the storage controller for writing to the storage device at the storage device address.
G06F 1/16 - Constructional details or arrangements
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
G06F 15/78 - Architectures of general purpose stored program computers comprising a single central processing unit
G06F 21/10 - Protecting distributed programs or content, e.g. vending or licensing of copyrighted material
G06F 21/64 - Protecting data integrity, e.g. using checksums, certificates or signatures
G06F 21/79 - Protecting specific internal or peripheral components, in which the protection of a component leads to protection of the entire computer to assure secure storage of data in semiconductor storage media, e.g. directly-addressable memories
G06T 19/00 - Manipulating 3D models or images for computer graphics
63.
Methods and apparatus for autocalibration of a wearable electrode sensor system
Methods and systems used in calibrating the position and/or orientation of a wearable device configured to be worn on a wrist or forearm of a user, the method comprises sensing a plurality of neuromuscular signals from the user using a plurality of sensors arranged on the wearable device, and providing the plurality of neuromuscular signals and/or signals derived from the plurality of neuromuscular signals as inputs to one or more trained autocalibration models, determining based, at least in part, on the output of the one or more trained autocalibration models, a current position and/or orientation of the wearable device on the user, and generating a control signal based, at least in part, on the current position and/or orientation of the wearable device on the user and the plurality of neuromuscular signals.
A61B 5/00 - Measuring for diagnostic purposes ; Identification of persons
A61B 5/06 - Devices, other than using radiation, for detecting or locating foreign bodies
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
G06F 3/0346 - Pointing devices displaced or positioned by the user; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
G06N 3/04 - Architecture, e.g. interconnection topology
64.
Auto-generating an artificial reality environment based on access to personal user content
Methods, systems, and storage media for auto-generating an artificial reality environment based on access to personal user content are disclosed. Exemplary implementations may: receive consent from a user to access user content on a user device, the user content comprising digital media; generate a user profile based at least in part on the user content; determine user preferences based at least in part on the user profile; generate an artificial reality environment based at least in part on the user preferences; and share the artificial reality environment with contacts of the user.
G06F 3/01 - Input arrangements or combined input and output arrangements for interaction between user and computer
G06F 3/04815 - Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
G06Q 50/00 - Systems or methods specially adapted for specific business sectors, e.g. utilities or tourism
65.
Systems and methods for spatial update latency compensation for head-tracked audio
A system can include a position sensor configured to output position data of a HWD. The system can include one or more processors configured to identify a first head angle of the HWD using the position sensor, generate an audio signal using the first head angle, identify a second head angle of the HWD using the position sensor, determine an angle error based at least on the first head angle and the second head angle, and apply at least one of a time difference or a level difference to the audio signal based at least on the angle error to adjust the audio signal. The system can include an audio output device configured to output the adjusted audio signal. By adjusting the audio signal using the angle error, the system can correct for long spatial update latencies and reduce the perceptual impact of such latencies for the user.
The disclosed system may include a support structure dimensioned for a user's hand. The system may also include transmitting electrodes coupled to a first finger portion of the support structure and may further include receiving electrodes coupled to a second, different finger portion of the support structure. The system may also include a controller that is coupled to the support structure and that is communicatively connected to the transmitting and receiving electrodes. The controller may also be configured to cause the transmitting electrodes to transmit a signal, detect at least some of the transmitted signal at the receiving electrodes and, based on the detected signal, determine that at least two fingers of the user's hand are touching each other. Various other methods, systems, and computer-readable media are also disclosed.
Disclosed herein are related to systems and methods for wireless communication. In one aspect, a system includes a first wireless interface configured to communicate at a first frequency band and a second frequency band. In one aspect, the system includes a second wireless interface configured to communicate at a third frequency band and a fourth frequency band. In one aspect, the system includes a switch configured to select either communication at the second frequency band or the third frequency band. In one aspect, the system includes a multi-band filter configured to couple i) the first wireless interface, ii) the second wireless interface, and iii) the switch, to a single antenna.
Disclosed herein are systems and methods related to using restricted target wake time in wireless communication. In one aspect, a first wireless communication device may configure a first field indicating (i) one or more traffic streams that are latency sensitive, and (ii) a direction of each of the one or more traffic streams between the first wireless communication device and a second wireless communication device. Each of the one or more traffic streams is to be communicated during a respective service period of a restricted target wake time (rTWT) schedule. The first wireless communication device may send a message including the first field to a second wireless communication device.
A sensor includes a plurality of pixels that each have dedicated compute circuitry within a compute layer. The plurality of pixels includes a first group of pixels and a second group of pixels. The first group of pixels is configured to detect light from a local area that has a first modulation frequency. The second group of pixels is configured to detect light from the local area that has a second modulation frequency. The compute layer is positioned below the plurality of pixels, and includes the compute circuitry for each of the plurality of pixels. The compute layer is configured to determine depth information for the local area using an indirect time-of-flight technique and one or both of the detected light that has the first modulation frequency and the detected light that has the second modulation frequency.
G01S 7/4915 - Time delay measurement, e.g. operational details for pixel components; Phase measurement
G01S 17/32 - Systems determining position data of a target for measuring distance only using transmission of continuous waves, whether amplitude-, frequency-, or phase-modulated, or unmodulated
A wrist device for gathering user data for biometric analysis and related methods, systems, and storage media are disclosed. The wrist device may include a low-power display, a base battery, and sensors. The base module may be configured to operate as a standalone module in a low-power mode. The base module may also be configured to gather data through the sensors. A computing module may be removably coupled with the base module. The computing module may include a high-definition display and a computing battery. The computing module may be configured to analyze the data when coupled to the base module. The computing battery of the computing module may be configured to charge the base battery of the base module when the computing module is coupled to the base module.
A system suppresses howl in a device including microphones and speakers, for example, an artificial reality headset. A speaker of the device presents audio content. The audio content presented by the speaker is received by a microphone of the device thereby creating a howl in certain situations. The system detects the presence of the howl in a region of the audio content using an adaptive notch filter. The system suppresses the howl by reducing gain of one or more frequencies of the audio content. The system may detect presence of the howl by monitoring flatness of the signal. The system may detect presence of the howl based on tonality detection based on linear prediction
A speaker produces acoustic frequencies within a housing that outputs the acoustic frequencies to a port. The produced frequencies travel through a cavity to the port which may have a cavity resonance that amplifies certain frequencies, affecting the frequency sensitivity of the speaker. To mitigate the cavity resonance, the speaker includes a membrane with regions having different breakup frequencies. One region is tuned to break up at a desired bandwidth of the speaker, and another region is tuned to break up at the cavity resonance, mitigating the distortion on frequency response due to the cavity resonance.
H04R 1/28 - Transducer mountings or enclosures designed for specific frequency response; Transducer enclosures modified by provision of mechanical or acoustic impedances, e.g. resonator, damping means
A leaky-mode acousto-optical modulator may be used to generate visual images suitable for direct viewing, without image-forming optics. To extend a field of view of the modulator to limits suitable for visual displays, the leaky-mode acousto-optical modulator may be equipped with a switchable beam redirector, e.g. a switchable-angle reflector, providing field of view portions one by one in a time-sequential manner. The field of view portions coalesce into a continuous synthetic field of view suitable for wide-angle visual display applications.
G03B 21/00 - Projectors or projection-type viewers; Accessories therefor
G02F 1/29 - Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the position or the direction of light beams, i.e. deflection
G02F 1/335 - Acousto-optical deflection devices having an optical waveguide structure
G09G 3/00 - Control arrangements or circuits, of interest only in connection with visual indicators other than cathode-ray tubes
The disclosed system may include a support structure that may be configured to house electronic components. The system may also include a first antenna mounted to the support structure. The first antenna may be configured to provide a wireless intralink on a first frequency to a local mobile electronic device. The system may also include multiple second antennas that are configured to established wireless interlinks on other frequencies to various external wireless networks. The second antennas may be positioned a specified minimum distance away from the first antenna and may be positioned at an at least partially opposing angle to each other. As such, the second antennas may provide at least a minimum threshold amount of spherical radiation to transmit and receive data using the established wireless interlinks. Various other apparatuses, methods of manufacturing, and mobile electronic devices are also disclosed.
A waveguide may include a substrate and an array of beam splitters embedded within the substrate, where each beam splitter within the array of beam splitters does not fully transect the substrate. Various other devices, systems, and methods of manufacture are also disclosed.
G02B 6/28 - Optical coupling means having data bus means, i.e. plural waveguides interconnected and providing an inherently bidirectional system by mixing and splitting signals
An eye tracking system with in-optical-assembly plane illumination is described. Side-emitting light emitting diodes (LEDs) aligned with a plane of an optical assembly of a near-eye display device are used to illuminate the eye of a user and generate glints that can be detected by an eye tracking camera. When a corrective optical lens or similar element is included in the optical assembly that may distort illumination beams from the light emitting diodes (LEDs), the distortion is mitigated by using in-package or externally modified LEDs that provide angled beams. In addition to in-package level mitigations such as reflectors or labels, edge portions of the distorting optical elements may be shaped or complemented with refractive elements to redirect the beams toward the eye.
Methods of assembling a head-mounted display (HMD) may include coupling a first digital projector assembly to an HMD frame, coupling a second digital projector assembly to the HMD frame, and then warping the HMD frame to optically align the first digital projector assembly with the second digital projector assembly. The warped HMD frame may be fixed such that the first digital projector assembly and the second digital projector assembly are optically aligned to within a predetermined threshold. Various other methods and systems are also disclosed.
An optical device includes a first electrode and a medium that includes ferroelectric liquid crystals and chiral dopants. The medium is located adjacent to the first electrode. The optical device may also include a second electrode distinct and separate from the first electrode. The optical device may be used as an optical dimming device, controlling an amount light passing through the optical device based on a voltage gradient provided to the optical device.
G02F 1/135 - Liquid crystal cells structurally associated with a photoconducting or a ferro-electric layer, the properties of which can be optically or electrically varied
G02F 1/137 - Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the intensity, phase, polarisation or colour based on liquid crystals, e.g. single liquid crystal display cells characterised by the electro-optical or magneto-optical effect, e.g. field-induced phase transition, orientation effect, guest-host interaction or dynamic scattering
79.
CONDUCTIVE ELASTOMERIC FOAM MATERIALS AND METHODS OF USE
Described herein are conductive elastomeric foam materials and methods of making and using the same. The conductive elastomeric foam materials include a polymeric matrix, one or more conductive fillers, and one or more foaming agents. The polymeric matrix can include a thermoset polymer or a thermoplastic polymer. Also described herein are methods of making conductive elastomeric foam materials. Further described herein are molded products including the conductive elastomeric foam materials as described herein and wearable devices including the molded products.
This disclosure describes various examples of a system which uses a multi-bank, multi-port shared memory system that may be implemented as part of a system on a chip. The shared memory system may have particular applicability in the context of an artificial reality system, and may be designed to have distributed or varied latency for one or more memory banks and/or one or more components or subsystems within the system on a chip. The described shared memory system may be logically a single entity, but physically may have multiple memory banks, each accessible by any of a number of components or subsystems. In some examples, the memory system may enable concurrent, common, and/or shared access to memory without requiring, in some situations, full locking or arbitration.
A technology to have better experiences, e.g., beyond-arm's-length experiences, with fine-tuned interactions in an extended-reality environment can include methods, extended-reality-compatible devices, and/or systems configured to generate, e.g., via an extended-reality-compatible device, a copy of a representation of an object in an extended-reality environment, to initiate control of the copy of the representation of the object according to a schema; and to control the copy of the representation of the object at a first location in the extended-reality environment, e.g., the first location including a location beyond-arm's-length distance from a second location in the extended-reality environment, the second location being a location of an avatar or representation of a user and/or a controller of the extended-reality-compatible device.
G06T 19/00 - Manipulating 3D models or images for computer graphics
G06F 3/04815 - Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
82.
TECHNOLOGY FOR CREATING, REPLICATING AND/OR CONTROLLING AVATARS IN EXTENDED REALITY
A technology for creating, replicating and/or controlling avatars in an extended-reality (ER) environment can include methods, extended-reality-compatible devices, and/or systems configured to generate, e.g., via an ER-compatible device, a copy of a representation of a user, e.g., a primary avatar, and/or an object in an ER environment, to initiate a recording of the graphical representation of the user or object according to a schema; to produce a copy of the recording of the graphical representation of the user as a new graphical representation of the user, e.g., a new avatar, in the ER environment; and to control the new graphical representation of the user at a first location in the ER environment. In examples, the technology enables moving the graphical representation of the user around the ER environment while the new graphical representation of the user performs motion and/or produces sound from the copy of the recording.
A method for requesting a recommendation for content items in a recommender system is provided. The method includes extracting a user attribute from the recommender system, identifying multiple content items that match a user preference based on the user attribute, determining an interest value for the user on each content item based on an affinity between the user attribute and a content item attribute, and providing the user a list of content items ranked according to the interest value for the user. A system including a memory storing instructions and a processor configured to execute the instructions and cause the system to perform the above method is also provided.
H04N 21/482 - End-user interface for program selection
H04N 21/258 - Client or end-user data management, e.g. managing client capabilities, user preferences or demographics or processing of multiple end-users preferences to derive collaborative data
A disclosed apparatus may include at least one touch sensor and an integrated film that includes at least one antenna, wherein the integrated film is at least partially disposed on the at least one touch sensor such that the at least one touch sensor operates as a radiating element for the at least one antenna. Various other apparatuses, systems, and methods are also disclosed.
A liquid lens with flexible transparent active layers on both sides of a fluid is transformed along two distinct deformation axes. The flexible transparent active layers include piezoelectric materials that actuate the lens in response to applied voltage(s). The piezoelectric properties and actuation mechanism of the transparent layers are arranged to deform the lens cylindrically along different axes resulting in a net spherical deformation or a combination of spherical and cylindrical deformation with substantially less distortion than spherically deforming layers. The piezoelectric active layers may be polymer or ceramic with isotropic or anisotropic mechanical stiffness. Alternatively, a pair of transparent, internal layers are positioned between the front and rear surfaces. The active layers, front and/or rear, are dual layers affixed together with an adhesive or single layers.
G02B 3/14 - Fluid-filled or evacuated lenses of variable focal length
G02B 1/06 - Optical elements characterised by the material of which they are made; Optical coatings for optical elements made of fluids in transparent cells
G02B 26/00 - Optical devices or arrangements for the control of light using movable or deformable optical elements
G02B 27/00 - Optical systems or apparatus not provided for by any of the groups ,
A device may one or more processors. The one or more processors may be configured to receive, via a transceiver from a first device, quality of service (QoS) information for traffic between the first device and an access point (AP). The one or more processors may be configured to wirelessly transmit, via the transceiver to the AP, a frame including address information of the first device and the QoS information.
An optical system is configured to support eye tracking operations in a head mounted device. The optical system includes a transparent layer, a laser, and first and second output couplers. The laser is configured to emit light into the transparent layer. The first output coupler is configured to out-couple a first portion of the light from a first location of the transparent layer. The second output coupler is configured to out-couple a second portion of the light from a second location of the transparent layer. The first and second output couplers are in the field-of-view of a user. A combination of the first portion of the light and the second portion of the light out-coupled from the first output coupler and the second output coupler is configured to generate a fringe pattern, for example, on an eye of a user of the head mounted device.
G01B 11/25 - Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures by projecting a pattern, e.g. moiré fringes, on the object
G02B 6/12 - Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings of the optical waveguide type of the integrated circuit kind
G02B 27/00 - Optical systems or apparatus not provided for by any of the groups ,
G06V 40/18 - Eye characteristics, e.g. of the iris
H01S 5/183 - Surface-emitting [SE] lasers, e.g. having both horizontal and vertical cavities having only vertical cavities, e.g. vertical cavity surface-emitting lasers [VCSEL]
H01S 5/40 - Arrangement of two or more semiconductor lasers, not provided for in groups
88.
OPTICAL IMAGE STABILIZATION WITH ASYMMETRIC STROKE FOR CAMERA DEVICES
Embodiments of the present disclosure relate to a camera device with optical image stabilization (OIS) having a range of motion that is asymmetric along two spatial dimensions. The camera device includes an image sensor, a lens assembly in an optical series with the image sensor, and an OIS assembly. The OIS assembly initiates a first motion of at least one of the image sensor and the lens assembly along a first direction parallel to a first axis, the first motion having a first range. The OIS assembly further initiates a second motion of at least one of the image sensor and the lens assembly along a second direction parallel to a second axis orthogonal to the first axis, the second motion having a second range different than the first range.
G03B 5/00 - Adjustment of optical system relative to image or object surface other than for focusing of general interest for cameras, projectors or printers
G02B 27/64 - Imaging systems using optical elements for stabilisation of the lateral and angular position of the image
G04G 21/02 - Detectors of external physical values, e.g. temperature
H04N 23/55 - Optical parts specially adapted for electronic image sensors; Mounting thereof
H04N 23/57 - Mechanical or electrical details of cameras or camera modules specially adapted for being embedded in other devices
H04N 23/68 - Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
89.
OPTICAL MODULATOR AND IMAGE PROJECTOR BASED ON LEAKY-MODE WAVEGUIDE WITH SPATIAL MULTIPLEXING
A leaky-mode acousto-optical modulator may be used to generate visual images suitable for direct viewing, without image-forming optics. To extend a field of view of the modulator to limits suitable for visual displays, an array of acoustic transducers is coupled to a waveguide along its length. Each acoustic transducer generates an acoustic wave causing a corresponding leaky mode to appear. A reflector is disposed proximate each transducer for directing the out-coupled light portion at a unique angle, providing a corresponding field of view portion for that direction. The field of view portions coalesce into a continuous synthetic field of view suitable for wide-angle visual display applications.
G02F 1/035 - Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the intensity, phase, polarisation or colour based on ceramics or electro-optical crystals, e.g. exhibiting Pockels or Kerr effect in an optical waveguide structure
G02F 1/00 - Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics
G02F 1/03 - Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the intensity, phase, polarisation or colour based on ceramics or electro-optical crystals, e.g. exhibiting Pockels or Kerr effect
90.
SYSTEM AND METHOD FOR PROVIDING SPATIOTEMPORAL VISUAL GUIDANCE WITHIN 360-DEGREE VIDEO
Features described herein generally relate to providing spatiotemporal guidance within a 360-degree video. Particularly, while a 360-degree video is being displayed to a user, a current location of the view of the user within the 360-degree video is determined. Additionally, regions of interest may be identified within the 360-degree video, along with their location and a time in which the regions of interest are active within the 360-degree video. A visual guide is then overlaid onto the 360-degree video that indicates a current location of the view of the user, as well as locations of one or more regions of interest, as well as the time during which each region of interest is active (able to be viewed) within the 360-degree video. By viewing the visual guide, the user may be able to anticipate a location and timing of regions of interest within the 360-degree video and adjust their gaze to the correct location at the correct time to view such regions of interest.
The disclosed system may include a support structure having a top portion and a first bottom portion. The system may also include at least one wireless communication device positioned on the top portion of the support structure. The system may further include at least one antenna positioned in the first bottom portion of the support structure. The antenna in the first bottom portion of the support structure may be electrically connected to the wireless communication device positioned on the top portion of the support structure. Various other wearable devices, apparatuses, and methods of manufacturing are also disclosed.
The disclosed devices and systems may include transparent antennas and touch displays that may be included in wearable devices. An example wearable device may include a display module including a display stack, the display stack including a display cover layer, an antenna layer, and a display layer, wherein the antenna layer (1) is positioned between the display cover layer and the display layer, and (2) includes at least one antenna element. Various other systems and devices are disclosed.
A lens system may include a lens having a liquid crystal layer and a driving electrode array disposed on a first side of the liquid crystal layer, the driving electrode array including a plurality of driving zones that are independently operable. The plurality of driving zones may each include a plurality of driving electrodes consecutively arranged along a radial direction extending from a central region of the lens. The lens system may include a controller that is configured to independently operate each of the plurality of driving zones via a separate plurality of bus lines coupled to each of the plurality of driving zones. The controller may be configured to operate at least one of the driving zones based on a gaze direction of the user. Various other devices, systems, and methods are also disclosed.
G02F 1/29 - Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the position or the direction of light beams, i.e. deflection
An apparatus, system, and method for a lens assembly for a head-mounted device includes a frame and a lens assembly. The lens assembly is configured to be carried by the frame. The lens assembly includes a lens bulk, an edge thickness, and a power layer. The edge thickness includes a predetermined thickness value that is configured to be compatible with (e.g., a slot in) the frame. The power layer is configured to be set to one of a number of predetermined power levels while maintaining the edge thickness within the predetermined thickness value. The power layer is configured to define a course power level of the lens assembly (e.g., in the range of −6.00 D to +6.00 D).
A curved battery cell may include (1) a positive terminal, (2) a negative terminal, (3) one or more curved electrodes, and (4) an outer case encasing the one or more curved electrodes that includes at least one curved surface having a non-uniform radius of curvature. A curved battery pack may include (1) a first curved battery cell with a first outer surface having a first non-uniform curvature and (2) a second curved battery cell with a second outer surface having a second non-uniform curvature. Various other apparatus, systems, and methods are also disclosed.
H01M 50/107 - Primary casings, jackets or wrappings of a single cell or a single battery characterised by their shape or physical structure having curved cross-section, e.g. round or elliptic
H01M 50/247 - Mountings; Secondary casings or frames; Racks, modules or packs; Suspension devices; Shock absorbers; Transport or carrying devices; Holders specially adapted for portable devices, e.g. mobile phones, computers, hand tools or pacemakers
The disclosed system may include a housing dimensioned to secure various components including at least one physical processor and various sensors. The system may also include a camera mounted to the housing, as well as physical memory with computer-executable instructions that, when executed by the physical processor, cause the physical processor to: acquire images of a surrounding environment using the camera mounted to the housing, identify features of the surrounding environment from the acquired images, generate a map using the features identified from the acquired images, access sensor data generated by the sensors, and determine a current pose of the system in the surrounding environment based on the features in the generated map and the accessed sensor data. Various other methods, apparatuses, and computer-readable media are also disclosed.
H04N 13/239 - Image signal generators using stereoscopic image cameras using two 2D image sensors having a relative position equal to or related to the interocular distance
H04N 23/54 - Mounting of pick-up tubes, electronic image sensors, deviation or focusing coils
97.
PSEUDOMONOPOLAR ELECTRODE CONFIGURATIONS FOR EMG SENSING
Disclosed herein are methods, systems, apparatuses, and media for sensing neuromuscular signals. In one example, a device for sensing neuromuscular signals comprises a wearable structure. The device includes a plurality of signal electrodes aligned along an interior portion of the wearable structure, each signal electrode configured to detect neuromuscular signals. The device further includes a plurality of amplifiers corresponding to the plurality of signal electrodes, wherein an amplifier has: a first input operatively coupled to a corresponding signal electrode; a second input; and an output corresponding to a neuromuscular signal channel. The device further includes circuitry configured to generate a common mode reference signal directly or indirectly based on signals from one or more electrodes, wherein the second input of each amplifier of the plurality of amplifiers is configured to receive the common mode reference signal or a signal based on the common mode reference signal.
An interposer equipped with a heat spreader and a multi-board system for an electronic device that includes an interposer equipped with a heat spreader between at least two boards. In examples, the interposer may include a heat spreader having an active portion and, optionally, a passive portion. In examples, the active portion may include a vapor chamber, heat pipe, or isothermal plate. In examples, the passive portion may thermally couple the active portion to a heat dissipation device such as a heat sink and/or an outer frame of the electronic device.
A camera module includes an image sensor die, high-density interconnect (HDI) tape, and a number of passive electronic components. The image sensor die has a first side and a second side. The first side includes a pixel array, and the second side includes a number of bonding pads. The HDI tape is a flexible substrate coupled to the image sensor. The HDI tape includes a number of traces positioned between a first side and a second side. The first side of the HDI tape is coupled to the second side of the image sensor die. The passive electronic components are coupled to the second side of the HDI tape and provide rigidity to the camera module.
Disclosed herein are related to devices and methods for communication. In one aspect, a device includes a first processor and a second processor. The first processor may be configured to generate a first set of packets associated with an application data unit in a first layer corresponding to content data. Each packet of the first set of packets may include a flag indicative of an association with the application data unit. The second processor may be configured to generate a second set of one or more packets in a second layer for transmission, in response to determining that the first set of packets is associated with the application data unit according to flags of the first set of packets. The second processor may be configured to schedule to transmit the second set of one or more packets in the second layer within a defined time period.