An image capture device may capture multiple audio content during capture of visual content. A viewing window for the visual content and rotational position of the image capture device during capture of the visual content may be used to generate modified audio content from the multiple audio content. The modified audio content may provide sound for playback of a punchout of the visual content using the viewing window.
H04N 21/43 - Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronizing decoder's clock; Client middleware
G06F 1/16 - Constructional details or arrangements
H04N 21/431 - Generation of visual interfaces; Content or additional data rendering
H04N 21/44 - Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to MPEG-4 scene graphs
H04N 21/472 - End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification or for manipulating displayed content
An image capture device may capture a video. A part of the video may include depiction of an object of interest, and another part of the video may not include depiction of the object of interest. Stabilization of the video may switch between object stabilization and movement stabilization based on whether or not the video depicts the object of interest. The part of the video depicting the object of interest may be stabilized using object stabilization. The part of the video not depicting the object of interest may be stabilized using movement stabilization.
An image capture device reduces noise floor using multiple microphones. The image capture device includes a processor that obtains a front microphone signal from a front microphone, the front microphone signal having a noisy noise floor portion, obtains a rear microphone signal from a rear microphone, sets a splice point based on mitigation of the noisy noise floor portion relative to a speech frequency range, and combines a substantially clean noise floor portion of the rear microphone signal at or below the splice point with a remaining portion of the front microphone signal above the splice point to generate a microphone signal.
H04R 1/40 - Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by combining a number of identical transducers
H04R 3/04 - Circuits for transducers for correcting frequency response
Adaptive acquisition control including limited luminance motion blur reduction for image and video acquisition and processing includes obtaining a target gain value and a target exposure duration value by obtaining a current scene luminance value, obtaining a current maximum gain using the current scene luminance value, obtaining a current interpolated gain-exposure duration curve using the current maximum gain and a current angular speed value, and obtaining the target gain value and the target exposure duration value from the current interpolated gain-exposure duration curve in accordance with a current maximum exposure duration. Adaptive acquisition control including limited luminance motion blur reduction for image and video acquisition and processing includes obtaining a frame of a video in accordance with the target scene exposition data and outputting the video.
Adaptive acquisition control timing control for image and video acquisition and processing includes obtaining, by an auto-exposure compensation component of an image processing pipeline of an image capture apparatus, during a current sensor readout window corresponding to obtaining current input image data from an image sensor of the image capture apparatus, in accordance with defined image acquisition and processing timing data, current auto-exposure compensation data for processing subsequent input image data obtained from the image sensor during a subsequent sensor readout window sequentially immediately subsequent to the current sensor readout window, and outputting, to a second data storage unit instance of the image processing pipeline, by the auto-exposure compensation component, during the current sensor readout window, the current auto-exposure compensation data.
Video frames of a video may be marked with visual patterns to identify individual video frames. The video may be changed by applying one or more effects to the video. The accuracy with which the changes were made to the video by the effect(s) may be determined using the visual patterns marked on the video frames.
G11B 27/36 - Monitoring, i.e. supervising the progress of recording or reproducing
G06K 19/06 - Record carriers for use with machines and with at least a part designed to carry digital markings characterised by the kind of the digital marking, e.g. shape, nature, code
Features to be enabled for an image capture device may be determined based on user subscription to a feature plan and/or user usage of the image capture device. The features for the image capture device may be enabled through firmware update or code unlock.
An image capture device detects a wind whistle using two or more microphones. The image capture device includes a processor that obtains microphone signals from the two or more microphones and measures coherence values between the microphone signals across a frequency band. The frequency band includes frequency bins, and the processor measures a coherence value for each frequency bin. Based on a detection of an elevated coherence value in a frequency bin, the processor determines the presence of a whistle. The processor attenuates the frequency bin based on a determination that the elevated coherence value is above a threshold.
G10K 11/178 - Methods or devices for protecting against, or for damping, noise or other acoustic waves in general using interference effects; Masking sound by electro-acoustically regenerating the original acoustic waves in anti-phase
Wind processing performance is improved by the placement of one or more microphones on a device. The device includes one or more microphones on the front surface, one or more microphones on the top surface, and one or more microphones on the rear surface. The rear surface is a heatsink, and the one or more microphones on the rear surface are located in recesses created by two or more fins of the heatsink. The device includes a processor that performs wind processing based on signals from the one or more microphones on the front surface, signals from the one or more microphones on the top surface, signals from the one or more microphones on the rear surface, or any combination thereof.
G03B 17/55 - APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR - Details of cameras or camera bodies; Accessories therefor with provision for heating or cooling, e.g. in aircraft
In one aspect of the present disclosure, a gimbal assembly is described for use with an image capturing device. The gimbal assembly includes a motor assembly, a first housing defining an internal compartment that is configured and dimensioned to receive the motor assembly, and a second housing that is mechanically connected to the motor assembly such that actuation of the motor assembly causes relative rotation between the first and second housings. The first housing includes a first guide that is configured and dimensioned to support transmission media adapted to communicate electrical and/or digital signals. The second housing defines a channel that is configured and dimensioned to receive the first guide such that the first guide extends into the second housing through the channel. The transmission media is supported on the first guide such that the first guide routes the transmission media from the first housing into the second housing.
H04N 23/68 - Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
F16M 11/10 - Means for attachment of apparatus; Means allowing adjustment of the apparatus relatively to the stand allowing pivoting around a horizontal axis
F16M 11/20 - Undercarriages with or without wheels
H04N 7/10 - Adaptations for transmission by electrical cable
H05K 5/00 - Casings, cabinets or drawers for electric apparatus
H05K 5/02 - Casings, cabinets or drawers for electric apparatus - Details
An image capture device is disclosed that includes: a front housing portion; a rear housing portion including a heatsink that is connected to the front housing portion; and a printed circuit board (PCB) module that is operatively connected to the front housing portion. The PCB module includes: a front PCB assembly; a rear PCB assembly; and a connector that extends between the front PCB assembly and the rear PCB assembly. The connector includes a flexible construction that facilitates reconfiguration of the PCB module during assembly of the image capture device between a first configuration, in which the front PCB assembly and the rear PCB assembly are oriented in non-parallel relation, and a second configuration, in which the front PCB assembly and the rear PCB assembly are oriented in generally parallel relation.
An image capture device includes a housing and a heatsink that forms an integral portion of the housing and draws heat from a heat generating component of the heatsink. The image capture device includes a circuit board enclosed by the housing and comprising a chip. The image capture device includes a gasket that connects the heatsink and the circuit board. The gasket encloses sides of the chip and the heatsink encloses a top of the chip so that electromagnetic interference does not affect other components of the image capture device.
The lens attachment includes a body that overlays a lens of an image capture device and one or more lights integrated with the body that emits light in a direction so that the light is not directed into the lens of the image capture device. The lens attachment includes a connector integrated with the body that secures the lens attachment against a housing of the image capture device.
Systems and methods are disclosed for high dynamic rate processing based on angular rate measurements. For example, methods may include receiving a short exposure image that was captured using an image sensor; receiving a long exposure image that was captured using the image sensor; receiving an angular rate measurement captured using an angular rate sensor attached to the image sensor during exposure of the long exposure image; determining, based on the angular rate measurement, whether to apply high dynamic range processing to an image portion of the short exposure image and the long exposure image; and responsive to a determination not to apply high dynamic range processing to the image portion, selecting the image portion of the short exposure image for use as the image portion of an output image and discard the image portion of the long exposure image.
H04N 23/741 - Circuitry for compensating brightness variation in the scene by increasing the dynamic range of the image compared to the dynamic range of the electronic image sensors
H04N 23/68 - Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
H04N 23/73 - Circuitry for compensating brightness variation in the scene by influencing the exposure time
A camera is configured for use with a removable camera lens cover, which can be secured to or removed from the camera by a user without the use of a tool set. The mechanism which allows the lens cover to be secured to and removed from the camera includes a set of wires embedded into the lens cover and a set of wedges protruding from the lens wall of the camera. To secure the lens cover to the camera, the lens cover is placed onto the front of the camera and rotated until the wires align with corresponding wedges, securing the wires underneath the tapered surface of the wedges. To remove the lens cover from the camera, a force is applied outward and normal to the lens cover, causing the wires to flex outward and enabling the rotation and removal of the lens cover from the camera.
G03B 17/12 - Bodies with means for supporting objectives, supplementary lenses, filters, masks, or turrets
G03B 17/55 - APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR - Details of cameras or camera bodies; Accessories therefor with provision for heating or cooling, e.g. in aircraft
22.
HEATSINKS AND THERMAL ARCHITECTURE OF AN IMAGE CAPTURE DEVICE
A image capture device including: heat generating devices, one or more batteries, and a sensor heat spreader. The heat generating devices generate a thermal load. The sensor heat spreader is in thermal communication with one or more of the heat generating devices and extends from the one or more of the heat generating devices to the one or more batteries so that all or a portion of the thermal load is transferred to the one or more batteries.
Auto exposure processing for spherical images improves image quality by reducing visible exposure level variation along a stitch line within a spherical image. An average global luminance value is determined based on luminance values determined for first and second images, which are based on auto exposure configurations of first and second image sensors used to obtain those first and second images. Delta luminance values are determined for the first and second images using the average global luminance value. The first and second images are then updated using the delta luminance values, and the updated first and second images are used to produce a spherical image.
H04N 23/45 - Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from two or more image sensors being of different type or operating in different modes, e.g. with a CMOS sensor for moving images in combination with a charge-coupled device [CCD] for still images
G06T 3/40 - Scaling of a whole image or part thereof
H04N 9/77 - Circuits for processing the brightness signal and the chrominance signal relative to each other, e.g. adjusting the phase of the brightness signal relative to the colour signal, correcting differential gain or differential phase
A camera housing includes a first surface and a second surface noncoplanar with the first surface. A first interconnect mechanism is coupled to the first surface and rotatable between a collapsed position and an extended position. In the collapsed position, protrusions of the first interconnect mechanism extend parallel to the first surface. In the extended position, the protrusions of the first interconnect mechanism extend in a perpendicular manner away from the first surface. A second interconnect mechanism is coupled to the second surface and rotatable between a collapsed position and an extended position. In the collapsed position, protrusions of the second interconnect mechanism include coplanar surfaces and extend adjacent to the second surface. In the extended position, the protrusions of the second interconnect mechanism extend in a perpendicular manner away from the second surface.
G03B 17/04 - Bodies collapsible, foldable or extensible, e.g. book type
G03B 17/12 - Bodies with means for supporting objectives, supplementary lenses, filters, masks, or turrets
G03B 17/55 - APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR - Details of cameras or camera bodies; Accessories therefor with provision for heating or cooling, e.g. in aircraft
A set of image portions is obtained. A pixel series that includes respective representative values of the image portions is obtained. Spectral data are obtained based on the pixel series. Significant local maxima are identified in the spectral data. An artificial light is determined to have been captured in the set of the image portions in response to determining that a number of the significant local maxima is greater than one. An anti-flicker feature of an image capture device is enabled. in response to determining that the artificial light is captured in the set of the image portions.
H04N 23/745 - Detection of flicker frequency or suppression of flicker wherein the flicker is caused by illumination, e.g. due to fluorescent tube illumination or pulsed LED illumination
H04N 23/71 - Circuitry for evaluating the brightness variation
28.
METHODS AND DEVICES FOR ACTIVE ATHERMALIZATION AND LENS POSITION INDEXING
Devices for active athermalization and lens position indexing include an image capture device that includes a lens assembly, a lens mount, a memory, a printed circuit board (PCB), an image sensor, a temperature sensor, and actuator, and a processor. The memory stores a calibration look up table (LUT) that includes focus positions across a temperature range. The image sensor may be disposed on the PCB. The temperature sensor measures a temperature of the lens assembly. The processor determines a position of the lens assembly relative to the image sensor to maintain a focus point over the temperature range based on the calibration LUT, the measured temperature, or both. The processor may be configured to transmit a control signal to the actuator to modify the position of the lens assembly relative to the image sensor to maintain the focus point based on the measured temperature.
H04N 23/52 - Elements optimising image sensor operation, e.g. for electromagnetic interference [EMI] protection or temperature control by heat transfer or cooling elements
H04N 17/00 - Diagnosis, testing or measuring for television systems or their details
H04N 23/68 - Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
H04N 23/55 - Optical parts specially adapted for electronic image sensors; Mounting thereof
H04N 23/54 - Mounting of pick-up tubes, electronic image sensors, deviation or focusing coils
Modular lens assemblies and mounting systems are disclosed. A modular lens assembly includes a removable portion and a fixed portion. The removable portion includes a first lens stack configured to produce a near-collimated ray path or a collimated ray path. The fixed portion includes a second lens stack configured to receive the near-collimated ray path or the collimated ray path from the removable portion. The modular lens assembly may be implemented in an image capture device. An image sensor of the image capture device is positioned at an end of the modular lens assembly. The image sensor is configured to capture images based on light incident on the image sensor through the first lens stack and the second lens stack such that the light incident on an outer lens of the first lens stack is refracted through the second lens stack to the image sensor.
An image capture device includes an image sensor and a processor. The image sensor is configured to capture a first plurality of frames, a second plurality of frames, and a third plurality of frames. The processor includes a first denoising layer and a second denoising layer. The first denoising layer includes a first denoiser, a second denoiser, and a third denoiser. The first denoiser is configured to denoise the first plurality of frames and output a first denoised frame. The second denoiser is configured to denoise the second plurality of frames and output a second denoised frame. The third denoiser is configured to denoise the third plurality of frames and output a third denoised frame. The second denoising layer includes a fourth denoiser. The fourth denoiser is configured to output a denoised frame based on the first denoised frame, the second denoised frame, and the third denoised frame.
Adaptive acquisition control includes obtaining a processed image by an image capture apparatus, which includes, a target exposure component that obtains a target exposure value in accordance with target exposure input data, an aggregate gain component that obtains a target aggregate gain value and a remaining gain value in accordance with aggregate gain input data, an auto-exposure compensation component that obtains an auto-exposure compensation tone curve in accordance with auto-exposure compensation input data, a contrast control component that obtains a contrast control tone curve and a contrast control black point value, and a tone control driver that obtains a tone control tone curve and a tone control black point value, and an image signal processor that processes the current input image in accordance with the tone control tone curve and the tone control black point value to produce the processed image.
H04N 23/72 - Combination of two or more compensation controls
H04N 23/71 - Circuitry for evaluating the brightness variation
H04N 23/741 - Circuitry for compensating brightness variation in the scene by increasing the dynamic range of the image compared to the dynamic range of the electronic image sensors
Features to be enabled for an image capture device may be determined based on user subscription to a feature plan and/or user usage of the image capture device. The features for the image capture device may be enabled through firmware update or code unlock.
A lens adapter is attached to an image capture device to lower the profile of the image capture device when mounted to a mounting device. The lens adapter includes a lens that refracts light incident from an outer lens of the lens adapter into collimated light. The lens adapter includes an optical device that alters the direction of the collimated light and directs the collimated light to an image capture device lens assembly.
A mount (coupler) is disclosed that is configured to connect an image capture device to an accessory. The mount includes a female component that is configured for connection to the accessory and which defines stops, and a male component that is configured for removable insertion into the female component. The male component includes: a body; arms that are supported by the body and which are configured for engagement with the stops to thereby connect the male component to the female component; and a power hub that is supported by the body and which is configured to receive a power cord along a longitudinal axis to thereby deliver power and/or data to the mount. The male component is configured for disconnection from the female component upon deflection of the arms beyond a threshold release distance that is less than a maximum transverse (e.g., horizontal) cross-sectional dimension defined by the power cord.
A first device includes a processor and instructions. The instructions, when executed by the processor, cause the processor to configure the first device to receive network credentials; accept a connection request from a second device to connect to the first device; obtain the network credentials from the second device; configure the first device to connect to a third device subsequent to obtaining the network credentials; and transmit a request to connect to the third device, where request includes the network credentials.
Systems and methods are disclosed for image signal processing. For example, methods may include receiving an image from an image sensor; applying a filter to the image to obtain a low-frequency component image and a high-frequency component image; determining a first enhanced image based on a weighted sum of the low-frequency component image and the high-frequency component image, where the high-frequency component image is weighted more than the low-frequency component image; determining a second enhanced image based on the first enhanced image and a tone mapping; and storing, displaying, or transmitting an output image based on the second enhanced image.
H04N 23/741 - Circuitry for compensating brightness variation in the scene by increasing the dynamic range of the image compared to the dynamic range of the electronic image sensors
Video information may define spherical video content having a duration. Spherical video content may define visual content viewable from a point of view as a function of progress through the spherical video content. Path information may define a path selection for the spherical video content. Path selection may include movement of a viewing window within the spherical video content. The viewing window may define extents of the visual content viewable from the point of view as the function of progress through the spherical video content. Time lapse parameter information may define at least two of a time portion of the duration, an image sampling rate, and a time lapse speed effect. A time lapse video may be generated based on the video information, the path information, and the time lapse parameter information.
A video may be presented with a timeline representation of the video. The timeline representation may enable a user to jump to different parts of the video. An action curve for the video may be presented along the timeline representation. The action curve may include values that represent different actions taken/experienced by an image capture device while capturing the video. The action curve may help the user identify/locate different parts of the video.
Systems, apparatus, and methods for piggyback camera calibration. Existing piggybacked capture techniques use a “beauty camera” and an “action camera” to capture raw footage. The user directly applies the EIS stabilization track of a piggybacked action camera to the cinematic footage to create desired stable footage. Unfortunately, since the action camera may have been slightly offset from the cinematic video camera, the EIS stabilization data will only roughly approximate the necessary corrections. In other words, the user must manually fine tune the corrections. The disclosed embodiments use a calibration sequence to estimate a physical offset between the beauty camera and the action camera. Then, the estimated physical offset can be used to calculate an offset camera orientation for stabilizing the beauty camera. The foregoing process can be performed in-the-field before actual capture. This allows the user to check their set-up and fix any issues before capturing the desired footage.
A video may be captured by an image capture device. A stabilized view of the video may be generated by using a punchout of the video. The shape and/or size of the punchout may be dynamically changed based on stabilization performance of the video. The shape and/or size of the punchout may be changed when the punchout is moving within the video. The shape and/or size of the punchout may not be changed when the punchout is not moving within the video.
A camera includes a housing that defines a camera housing cavity and a lens assembly partially disposed in the camera housing cavity. The lens assembly comprises a lens, a lens cover, and a lens cavity that is defined between the lens and the lens cover. The camera includes a fluid path that extends from the lens cavity, through the camera housing cavity, and to an external environment. The camera includes a seal that partially obstructs the fluid path and is configured to equalize a humidity in the lens cavity with a humidity in the external environment.
H04N 23/52 - Elements optimising image sensor operation, e.g. for electromagnetic interference [EMI] protection or temperature control by heat transfer or cooling elements
H04N 23/55 - Optical parts specially adapted for electronic image sensors; Mounting thereof
45.
SYSTEMS, APPARATUS, AND METHODS FOR STABILIZATION AND BLENDING OF EXPOSURES
Systems, apparatus, and methods for stabilization and blending of exposures. So-called Electronic Image Stabilization (EIS) techniques use image manipulation software to compensate for camera motion. Unfortunately, existing EIS techniques cannot compensate for artifacts introduced by low shutter speed (e.g., blurs). Various embodiments of the present disclosure generate stabilized images from multiple exposures. In one exemplary embodiment, the stabilized exposures are blended using a linear sum of the color data for each pixel of the image. By stabilizing each exposure and linearly summing the light information, the camera shake can be removed, and the scene motion blur can be controlled. The stabilization and blending techniques enable a mathematical emulation of a selected shutter angle from many high-speed exposures.
Video frames of a video may be marked with visual patterns to identify individual video frames. The video may be changed by applying one or more effects to the video. The accuracy with which the changes were made to the video by the effect(s) may be determined using the visual patterns marked on the video frames.
G11B 27/031 - Electronic editing of digitised analogue information signals, e.g. audio or video signals
G11B 27/36 - Monitoring, i.e. supervising the progress of recording or reproducing
G06K 19/06 - Record carriers for use with machines and with at least a part designed to carry digital markings characterised by the kind of the digital marking, e.g. shape, nature, code
An image capture device may capture a video. A part of the video may include depiction of an object of interest, and another part of the video may not include depiction of the object of interest. Stabilization of the video may switch between object stabilization and movement stabilization based on whether or not the video depicts the object of interest. The part of the video depicting the object of interest may be stabilized using object stabilization. The part of the video not depicting the object of interest may be stabilized using movement stabilization.
09 - Scientific and electric apparatus and instruments
Goods & Services
Protective helmets; visors for helmets; bags specially adapted for protective helmets and sports helmets; helmet safety lights; protective helmets incorporating electronics; protective helmets incorporating electronic devices; Protective helmets for sports and motorcycle helmets featuring in-built communications, audio-visual and global positioning system software and apparatus; bicycle helmets; Helmets for motorcyclists; Motorcycle helmets; Motorcycle goggles; Protective helmets for cyclists; Protective helmets for sports; Protective sports helmets; Crash helmets; Crash helmets for cyclists; Helmets for use in sports; Protective helmets for motor cyclists; Sports helmets; Snowboarding helmets; Ski helmets; Riding helmets; Head guards for sports; Articles of protective clothing for wear by motorcyclists for protection against accident or injury; Helmets incorporating video cameras, namely, motorcycle helmets incorporating video cameras, Helmets incorporating Global Positioning Systems, namely, Motorcycle helmets incorporating GPS transceivers, Protective sports helmets incorporating GPS transceivers; Helmets incorporating audio and video equipment, namely, Motorcycle helmets incorporating apparatus for broadcasting, recording, transmission or reproduction of sound or images, Protective sports helmets incorporating apparatus for broadcasting, recording, transmission or reproduction of sound or images; Helmets incorporating barometers, altimeters, and gyroscopes, namely, Motorcycle helmets incorporating barometers, altimeters, and gyroscopes; Helmets incorporating tracking equipment and software, namely, motorcycle helmets incorporating GPS tracking devices and software for operating and controlling GPS tracking devices, Protective sports helmets incorporating GPS tracking devices and software for operating and controlling GPS tracking devices; Motion sensors and tracking sensors for use with helmets, namely, GPS tracking devices for use with protective helmets, motion detectors for use with protective helmets, and sensors for determining position, velocity and acceleration for use with protective helmets; Cameras for use with protective helmets; imaging equipment for use with helmets, namely, Video displays mounted in protective helmets, Eye pieces for helmet mounted displays used in protective helmets, Fixed and helmet mounted transparent electronic displays for providing aircraft crew members with navigational and operational information; Helmets with wireless and Internet connectivity, namely, Motorcycle helmets incorporating wireless transmitters and receivers, Protective sports helmets incorporating wireless transmitters and receivers; downloadable computer software for use in determining position, velocity, acceleration, motion, atmospheric pressure, distance, direction, altitude, and temperature of a helmet, and for wireless communication of audio and video, in the field of sports; downloadable software for controlling and operating electronic devices embedded in or attached to protective helmets
A camera mode to use for capturing an image or video is selected by estimating high dynamic range (HDR), motion, and light intensity with respect to a scene of the image or video to capture. An image capture device includes a HDR estimation unit to detect whether HDR is present in a scene of an image to capture, a motion estimation unit to determine whether motion is detected within the scene, and a light intensity estimation unit to determine whether a scene luminance for the scene meets a threshold. A mode selection unit selects a camera mode to use for capturing the image based on output of the HDR estimation unit, the motion estimation unit, and the light intensity estimation unit. An image sensor captures the image according to the selected camera mode.
H04N 23/667 - Camera operation mode switching, e.g. between still and video, sport and normal or high and low resolution modes
H04N 23/71 - Circuitry for evaluating the brightness variation
H04N 23/68 - Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
H04N 23/741 - Circuitry for compensating brightness variation in the scene by increasing the dynamic range of the image compared to the dynamic range of the electronic image sensors
Heating of an image capture apparatus battery to an operational temperature using an attached heater device is described. The image capture apparatus includes an image capture apparatus battery configured to functionally operate at a first operational temperature. The heater device includes a heater device battery configured to operate at a second operational temperature lower than the first operational temperature. A heater controller is implemented in one of the image capture apparatus or the heater device. The heater controller is configured to initiate a heating process based on a first defined event, set a current level of the heater device, maintain or adjust the current level of the heater device until a second defined event, and display an indication that the image capture apparatus can detect, capture, or record an image based on the image capture apparatus battery attaining the first operational temperature.
An image capture device performs audio processing based on a detection of whether a wind sock is attached to the image capture device. When a wind sock is detected, the image capture device performs stereo processing. When a wind sock is not detected, the image capture device performs wind processing.
An image capture device includes a housing, a processor, and three or more microphones. The housing includes a forward wall including a sensor, a rearward wall located opposite the forward wall, and a top wall connecting the forward wall and the rearward wall. The three or more microphones are configured to capture sound. The three or more microphones include a first microphone, a second microphone, and a third microphone. The processor is configured to receive the sound from the three or more microphones and to estimate a direction of arrival, reduce or remove wind noise, perform beamforming, or a combination thereof. The first microphone is located on or within the forward wall, the second microphone is located on or within the top wall, and the third microphone is located on or within the top wall and spaced apart from the second microphone.
H04R 1/40 - Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by combining a number of identical transducers
Methods and apparatus for shared image processing among multiple devices. In one embodiment, an exemplary action camera performs a partial multiband blend. Even though the action camera may not have resources to handle the multiband blend of the entire action camera's footage, it can do a significant portion. The partially blended content can be used in ready-to-share applications, or completely blended by another device.
Video information may define video content. The video content may be characterized by capture information. A remote device may transmit at least a portion of the capture information to a computing device. The computing device may identify one or more portions of the video content based on the transmitted capture information. The remote device may receive the identification of the identified portion(s) of the video content from the computing device. The remote device may stream one or more portions of the video information defining at least some of the identified portion(s) of the video content to the computing device. The streamed video information may enable the computing devices to provide a presentation of at least some of the identified portion(s) of the video content using a buffer of the streamed video information, which may allow the computing device to present the video content without permanently storing the video information.
H04N 21/472 - End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification or for manipulating displayed content
G11B 27/031 - Electronic editing of digitised analogue information signals, e.g. audio or video signals
An image capture device is described that includes a body; a mounting structure that is connected to the body; an integrated sensor-lens assembly (ISLA) that defines an optical axis and extends through the body and the mounting structure; and an accessory that is releasably connectable to the mounting structure via rotation through a range of motion less than approximately 90 degrees. The mounting structure and the accessory include corresponding angled bearing surfaces that are configured for engagement such that rotation of the accessory relative to the mounting structure creates a bearing effect that displaces the accessory along the optical axis to thereby reduce any axial force required during connection and disconnection of the accessory.
Methods and apparatus for seamlessly transitioning between lens projections. In one exemplary embodiment, a piecewise lens projection is composed of three (3) functions: (i) a first polynomial-based lens projection, (ii) a second “joining” lens projection, and (iii) a trigonometric lens projection. The piecewise lens projection characterizes virtualized lens distortion as a function of FOV; image data can be dynamically projected based on the virtualized lens distortion, regardless of FOV. In this manner, a user may achieve the visually familiar effects associated with a first lens definition for a first FOV, while still smoothly animating transitions to other lens projections (e.g., a larger FOV using stereographic projections).
A system includes a barrel mount disposed within a body, and the barrel mount includes a central axis. The system includes a lens barrel secured within the barrel mount and aligned with the central axis. The system includes an actuator that adjusts a position of the lens barrel by moving the barrel mount along the central axis. The system includes an actuator locking mechanism securable over the barrel mount and/or the lens barrel that prevents movement of the actuator and/or barrel mount when applying a force along the central axis towards the actuator.
An image capture device may include a touchscreen display, which may be used to receive user input. The image capture device may determine whether it is under water based on analysis of visual content captured by the image capture device. Responsive to determination that it is under water, the image capture device may change operation with respect to the touchscreen display.
A camera case that includes a buoyant material. The buoyant material forms a cavity defined by a top side, a right side, a bottom side, and a left side, and the cavity is configured to removably receive a camera. A front opening extends into the cavity. A rear opening is located opposite the front opening, and the rear opening is larger than the front opening so that the camera is extendable into the cavity through the rear opening. A button recess in the top side is aligned with a physical input interface of the camera. A mount opening extends through the bottom side of the camera case to receives a portion of a support member so that the support member and the camera are connectable through the mount opening. The camera case includes a sufficient amount of the buoyant material to float the camera in water when the camera is inserted into the cavity.
Video and corresponding metadata is accessed. Events of interest within the video are identified based on the corresponding metadata, and best scenes are identified based on the identified events of interest. A video summary can be generated including one or more of the identified best scenes. The video summary can be generated using a video summary template with slots corresponding to video clips selected from among sets of candidate video clips. Best scenes can also be identified by receiving an indication of an event of interest within video from a user during the capture of the video. Metadata patterns representing activities identified within video clips can be identified within other videos, which can subsequently be associated with the identified activities.
G11B 27/30 - Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on the same track as the main recording
G11B 27/031 - Electronic editing of digitised analogue information signals, e.g. audio or video signals
G10L 25/57 - Speech or voice analysis techniques not restricted to a single one of groups specially adapted for particular use for comparison or discrimination for processing of video signals
H04N 9/82 - Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
G11B 27/28 - Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
G06V 20/40 - Scenes; Scene-specific elements in video content
G10L 15/06 - Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
G10L 15/22 - Procedures used during a speech recognition process, e.g. man-machine dialog
H04N 5/77 - Interface circuits between an apparatus for recording and another apparatus between a recording apparatus and a television camera
G11B 27/13 - Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information not detectable on the record carrier the information being derived from movement of the record carrier, e.g. using tachometer
An image capture device is disclosed that includes a body and a door assembly that is configured for removable connection to the body. The door assembly includes a door body; a locking mechanism that is slidable in relation to the door body between a locked position and an unlocked position; and at least one biasing member that is configured for engagement (contact) with the door body and the locking mechanism to automatically move the locking mechanism into the locked position upon closure of the door assembly.
Image captured through a non-rectilinear lens may exhibit distortions. The distortions may be reduced by warping the image. However, warping the image may degrade the fidelity of the image. The warped image may be enhanced to increase the fidelity of the image. The enhancement may be applied to the portions of the image that were degraded from the warping.
A camera system captures an image in a source aspect ratio and applies a transformation to the input image to scale and warp the input image to generate an output image having a target aspect ratio different than the source aspect ratio. The output image has the same field of view as the input image, maintains image resolution, and limits distortion to levels that do not substantially affect the viewing experience. In one embodiment, the output image is non-linearly warped relative to the input image such that a distortion in the output image relative to the input image is greater in a corner region of the output image than a center region of the output image.
H04N 5/262 - Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects
H04N 21/2343 - Processing of video elementary streams, e.g. splicing of video streams or manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
H04N 23/80 - Camera processing pipelines; Components thereof
G06T 3/00 - Geometric image transformation in the plane of the image
G09G 5/00 - Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
G06T 3/40 - Scaling of a whole image or part thereof
G06T 3/20 - Linear translation of a whole image or part thereof, e.g. panning
H04N 21/4402 - Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to MPEG-4 scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
65.
Methods And Apparatus For Optimized Stitching Of Overcapture Content
Apparatus and methods for optimized stitching of overcapture content. In one embodiment, the optimized stitching of the overcapture content includes capturing the overcapture content; producing overlap bands associated with the captured overcapture content; downsampling the produced overlap bands; generating derivative images from the downsampled overlap bands; generating a cost map associated with the generated derivative images; determining shortest path information for the generated cost map; generating a warp file based on the determined shortest path information, the generated warp file being utilized for the optimized stitching of the overcapture content. Camera apparatus and a non-transitory computer-readable apparatus are also disclosed.
G06T 3/00 - Geometric image transformation in the plane of the image
H04N 7/08 - Systems for the simultaneous or sequential transmission of more than one television signal, e.g. additional information signals, the signals occupying wholly or partially the same frequency band
H04N 23/698 - Control of cameras or camera modules for achieving an enlarged field of view, e.g. panoramic image capture
66.
SYSTEMS AND METHODS FOR STABILIZING VIEWS OF VIDEOS
A viewing direction may define an angle/visual portion of a spherical video at which a viewing window is directed. A trajectory of viewing direction may include changes in viewing directions for playback of spherical video. Abrupt changes in the viewing directions may result in jerky or shaky views of the spherical video. Changes in the viewing directions may be stabilized to provide stabilized views of the spherical video. Amount of stabilization may be limited by a margin constraint.
H04N 13/167 - Synchronising or controlling image signals
H04N 13/117 - Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation the virtual viewpoint locations being selected by the viewers or determined by viewer tracking
67.
VIRTUAL WALL MAPPING FOR AERIAL VEHICLE NAVIGATION
An unmanned aerial vehicle (“UAV”), the UAV includes an electronic speed controller and a flight controller. The electric speed controller is interfaced with thrust motors of the UAV. The flight controller configured to: determine a geographic location and a velocity of the UAV, the velocity includes a first component and a second component. The flight controller is configured to determine a distance between the geographic location of the UAV and a closest segment of a no-fly zone. The flight controller is configured to determine a zone of deceleration, the zone of deceleration comprising: a distal section and a proximal section. The flight controller in response to the UAV crossing a switch point, located at an intersection of the distal section and the proximal section, changing a deceleration rate of the UAV from a first deceleration rate to a second deceleration rate by adjusting the electric speed controller and the thrust motors.
A gimbal mount system is configured to a couple to a gimbal coupled to and securing a camera. The gimbal mount system includes a handle, a power source, a user interface, a mounting interface, a communication interface, and a communication bus. The mounting interface is located within an end of the gimbal mount system and includes an opening configured to receive a reciprocal mounting protrusion of the gimbal. A locking mechanism removably couples the gimbal to the gimbal mount system. The communication interface is located within the mounting interface and is configured to couple to a reciprocal communication interface of the gimbal. The communication bus is coupled to the power source, user interface, and communication interface and is configured to provide power from the power source to the gimbal. The communication bus may provide instructions to the gimbal based on user input received via the user interface.
F16M 11/12 - Means for attachment of apparatus; Means allowing adjustment of the apparatus relatively to the stand allowing pivoting in more than one direction
F16M 11/16 - Heads - Details concerning attachment of head-supporting legs, with or without actuation of locking members therefor
F16M 11/18 - Heads with mechanism for moving the apparatus relatively to the stand
G03B 17/38 - Releasing-devices separate from shutter
69.
SYSTEMS AND METHODS FOR SWITCHING BETWEEN VIDEO VIEWS
A graphical user interface may provide presentation of a video edit during video edit playback. The video edit may include a segment of a video. When a user modifies the segment of the video included in the video edit, the graphical user interface may automatically switch the presentation of the video edit into presentation of the original video.
Implementations of a mobile platform for device control may use voice recognition. A user may use a mobile platform, such as a mobile application, on a mobile device to interpret and relay voice commands to a device. Voice recognition services may be integrated into the mobile application, the mobile device operating system (OS), or both.
Systems and methods are disclosed for circular stitching of images. For example, methods may include accessing a first image captured using a first image sensor; accessing a second image captured using a second image sensor; determining a cost table for a circular stitching boundary that includes overlapping regions of the first image and the second image; determining an extended disparity profile based on a periodic extension of the cost table and a smoothness criterion, wherein the extended disparity profile has a length greater than the width of the cost table; determining a binocular disparity profile of a length equal to the width of the cost table based on a contiguous subsequence of the extended disparity profile; and stitching the first image and the second image using the binocular disparity profile to obtain a combined image.
H04N 13/239 - Image signal generators using stereoscopic image cameras using two 2D image sensors having a relative position equal to or related to the interocular distance
G06T 7/593 - Depth or shape recovery from multiple images from stereo images
G06T 3/40 - Scaling of a whole image or part thereof
H04N 23/45 - Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from two or more image sensors being of different type or operating in different modes, e.g. with a CMOS sensor for moving images in combination with a charge-coupled device [CCD] for still images
Field variable tone mapping is performed for 360 content. An image capture device includes an image sensor and a processor. The image sensor obtains a hyper-hemispherical image and the processor performs local tone mapping (LTM) on a first area of the hyper-hemispherical image and performs global tone mapping (GTM) on a second area of the hyper-hemispherical image to obtain a processed image. The processor may be configured to display, store, output, or transmit the processed image.
H04N 23/741 - Circuitry for compensating brightness variation in the scene by increasing the dynamic range of the image compared to the dynamic range of the electronic image sensors
G06T 5/20 - Image enhancement or restoration by the use of local operators
A graphical user interface may switch between presentation of a timeline representation of media items and a tile representation of media items. The timeline representation may include graphical representation of the lengths of the media items. The tile representation may include graphical representation of the content of the media items.
G06F 3/0484 - Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
A video including visual content may be captured by an image capture device in motion. Stabilization performance information for the visual content may be determined. The stabilization performance information may characterize an extent to which desired stabilization is able to be performed using the visual content. The stabilization for the visual content may be changed based on the stabilization performance information.
Video content may be captured by an image capture device during a capture duration. The video content may include video frames that define visual content viewable as a function of progress through a progress length of the video content. Rotational position information may characterize rotational positions of the image capture device during the capture duration. Time-lapse video frames may be determined from the video frames of the video content based on a spatiotemporal metric. The spatiotemporal metric may characterize spatial smoothness and temporal regularity of the time-lapse video frames. The spatial smoothness may be determined based on the rotational positions of the image capture device corresponding to the time-lapse video frames, and the temporal regularity may be determined based on moments corresponding to the time-lapse video frames. Time-lapse video content may be generated based on the time-lapse video frames.
G11B 27/13 - Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information not detectable on the record carrier the information being derived from movement of the record carrier, e.g. using tachometer
Apparatus and methods for the stitch zone calculation of a generated projection of a spherical image. In one embodiment, a non-transitory computer-readable apparatus comprising a storage apparatus, the storage apparatus comprising instructions configured to, when executed by a processor apparatus, cause a computerized apparatus to identify a stitch line associated with an equatorial area of a plurality of spherical images; re-orient the plurality of spherical images in accordance with the stitch line; and project the re-oriented plurality of spherical images to a selected image projection type.
Systems, apparatus, and methods for motion transfer between different media. Two videos shot simultaneously are combined with aspects of each used to produce a higher-resolution frame interpolated video than would otherwise be possible by the devices alone. The interpolated video may be used in slow motion footage, a higher frame rate (e.g., virtual reality) application, or to add a realistic motion blur effect in post-processing (rather than in capture). A post-processing device may receive high-speed video from a high-speed camera and a low-speed video from a low-speed camera. The post-processing device may use motion data from the high-speed video to interpolate frames of (high-resolution) low-speed video producing low-speed interpolated video.
An image head with a housing, an integrated sensor and lens assembly (ISLA), a port, and internal componentry. The housing has a front side; a rear side; a top side; and a bottom side. The ISLA extends from the front side and is configured to detect images. The port is located on the rear side that is configured to electrically connect the image head to a base when the image head is inserted into the base. The internal componentry includes: a first microphone located within the front side below the ISLA; a second microphone located within the top side of the housing; a speaker located at the bottom side of the housing; a printed circuit board in communication with the integrated sensor of the ISLA; and a memory located on the printed circuit board and configured to store the images.
H04N 5/335 - Transforming light or analogous information into electric information using solid-state image sensors [SSIS]
H04N 23/55 - Optical parts specially adapted for electronic image sensors; Mounting thereof
H04N 23/54 - Mounting of pick-up tubes, electronic image sensors, deviation or focusing coils
H04N 23/663 - Remote control of cameras or camera parts, e.g. by remote control devices for controlling interchangeable camera parts based on electronic image sensor signals
A camera system with a camera body, a loudspeaker grille, a loudspeaker assembly, a waterproof membrane, and a support structure. The loudspeaker grille allows passage of sound waves from inside to outside the camera body. The loudspeaker assembly has a loudspeaker that converts a signal into the sound waves; and an audio circuit board that processes an audio signal into the signal. The waterproof membrane prevents moisture from passing into the camera body while allowing the sound waves to pass through. The support structure is located between the waterproof membrane and the loudspeaker assembly. A compressible spacer is coupled to the loudspeaker assembly. A loudspeaker adhesive connects the compressible spacer and the loudspeaker. A loudspeaker housing is configured to supporting the loudspeaker assembly, the loudspeaker grille, the waterproof membrane, and the support structure. A first loudspeaker cavity exists between the loudspeaker grille and the waterproof membrane.
The present teachings provide a camera housing that includes a first housing portion, a second housing portion, a hinge, a latch mechanism, and protrusions. The first housing portion forms a first portion of a cavity. The second housing portion forms a second portion of the cavity The hinge is located within the first housing portion so that the first housing portion is divided into a first piece and a second piece. The hinge movably connects the first housing portion to the second housing portion so that the second housing portion is movable relative to the first housing. The latch mechanism is movable between an open position and a closed position. The protrusion extend from a third housing portion, and the third housing portion is located opposite the first housing portion with respect to the cavity.
An image capture system and methods for auto-recording media data are herein disclosed. A method includes selecting an activity-specific monitor based on an activity type. The activity-specific monitor defines one or more auto-recording conditions that, when satisfied, cause the image capture system to record data. The auto-recording conditions are based on an audio profile, a video profile, a motion profile, or any combination thereof. The auto-recording conditions include one or more scene descriptions that correspond to the activity type.
G06F 21/53 - Monitoring users, programs or devices to maintain the integrity of platforms, e.g. of processors, firmware or operating systems during program execution, e.g. stack integrity, buffer overflow or preventing unwanted data erasure by executing in a restricted environment, e.g. sandbox or secure virtual machine
G06F 3/048 - Interaction techniques based on graphical user interfaces [GUI]
Apparatus and methods for stitching images, or re-stitching previously stitched images. Specifically, the disclosed systems in one implementation save stitching information and/or original overlap source data during an original stitching process. During subsequent retrieval, rendering, and/or display of the stitched images, the originally stitched image can be flexibly augmented, and/or re-stitched to improve the original stitch quality. Practical applications of the disclosed solutions enable, among other things, a user to create and stitch a wide field of view (FOV) panorama from multiple source images on a device with limited processing capability (such as a mobile phone or other capture device). Moreover, post-processing stitching allows for the user to convert from one image projection to another without fidelity loss (or with an acceptable level of loss).
Positions of an image capture device during capture of a video may be transferred to a computing device before the video is transferred to the computing device. The positions of the image capture device may be used to determine a viewing window for the video before the video is obtained. The viewing window may be used to present a stabilized view of the video when the video is obtained. For example, a stabilized view of the video may be presented as the video is streamed to the computing device.
The present teachings provide a system and method. The system and method include receiving images or video frames at a wireless receiver interface from a wireless transmitter. The system and method include performing decoder nudging while decoding the images or the video frames received by the wireless transmitter. Overclocking a display of a controller to an overclocked frequency. Outputting decoded images or decoded video frames to the display of the controller at the overclocked frequency.
H04N 7/18 - Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
B64C 39/02 - Aircraft not otherwise provided for characterised by special use
H04N 21/4402 - Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to MPEG-4 scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
H04N 21/4363 - Adapting the video stream to a specific local network, e.g. a IEEE 1394 or Bluetooth® network
Apparatus and methods for the non-uniform downsampling of captured panoramic images. In one embodiment, a computing device is disclosed that includes a processing apparatus and a non-transitory computer readable apparatus comprising a storage medium have one or more instructions stored thereon. The one or more instructions, when executed by the processing apparatus, being configured to: receive captured images, the captured images obtained using two or more image sensors; non-uniformly downsample the received captured images; and encode the non-uniformly downsampled images. In some variants, the non-uniformly downsampled images take into account a desired area of interest within the captured images. In some implementations, the computing device includes an image capture device. Methods and non-transitory computer readable apparatus are also disclosed.
After a command to stop recording a video is received, an image capture device may buffer footage in a buffer memory. The buffer memory may be used as a post-capture cache. The footage buffered in the buffer memory may be appended to the end of previously captured footage, appended to the beginning of subsequently captured footage, and/or used to bridge two separately captured footage.
A method includes determining an altitude of a camera of an aerial vehicle, determining a field of view (FOV) of a camera, generating a localized map, determining a relative position of the aerial vehicle on the localized map, and determining a relative heading of the aerial vehicle.
The image capture device includes a receptacle defined within a housing and an optic system that is removably connect with the receptacle and to generate thermal energy. The image capture device includes a heatsink positioned at a base of the receptacle. The heatsink includes an inner support positioned within the housing and an outer support in thermal communication with the optic system and the inner support. The heatsink further includes a gasket integrated with the receptacle to form a watertight seal, sandwiched between portions of the inner and outer supports, and flexibly retaining a physical connection between the optic system and the outer support.
G03B 17/55 - APPARATUS OR ARRANGEMENTS FOR TAKING PHOTOGRAPHS OR FOR PROJECTING OR VIEWING THEM; APPARATUS OR ARRANGEMENTS EMPLOYING ANALOGOUS TECHNIQUES USING WAVES OTHER THAN OPTICAL WAVES; ACCESSORIES THEREFOR - Details of cameras or camera bodies; Accessories therefor with provision for heating or cooling, e.g. in aircraft
The image capture device includes a receptacle defined within a housing and an optic system that is removably connected with the receptacle and configured to generate thermal energy. The image capture device includes a heatsink positioned at a base of the receptacle. The heatsink includes an inner support positioned within the housing and an outer support in thermal communication with the optic system and the inner support. The heatsink further includes a gasket integrated with the receptacle to form a watertight seal, sandwiched between portions of the inner and outer supports, and configured to flexibly retain a physical connection between the optic system and the outer support.
Images with an optical field of view are captured by an image capture device. An observed trajectory of the image capture device reflects the positions of the image capture device at different moments may be determined. A capture trajectory of the image capture device reflects virtual positions of the image capture device from which video content may be generated. The capture trajectory is determined based on a subsequent portion of the observed trajectory such that a portion of the capture trajectory corresponding to a portion of the observed trajectory is determined based on a subsequent portion of the observed trajectory. Orientations of punch-outs for the images are determined based on the capture trajectory. Video content is generated based on visual content of the images within the punch-outs.
In a video capture system, a virtual lens is simulated when applying a crop or zoom effect to an input video. An input video frame is received from the input video that has a first field of view and an input lens distortion. A selection of a sub-frame representing a portion of the input video frame is obtained that has a second field of view smaller than the first field of view. The sub-frame is processed to remap the input lens distortion to a desired lens distortion in the sub-frame. The processed sub-frame is output.
A method and apparatus for night lapse video capture are disclosed. The apparatus includes an image sensor, a processor, a denoiser, and an encoder. The image sensor is configured to capture image data. The processor is configured to process the image data using an image processing pipeline. The denoiser is configured to perform denoising based on a motion estimation that is based on a comparison of a portion of the first image and a portion of the second image. The encoder is configured to encode the denoised image in a video format. The encoder is configured to store a video file in the video format.
Depiction of a physical whiteboard may be captured by an image capture device. The depiction of the physical whiteboard may be detected and used to generate a virtual whiteboard. A change in the depiction of the physical whiteboard may be detected and used to change the virtual whiteboard. A virtual change to the virtual whiteboard may be received and used to change the virtual whiteboard. The change to the virtual whiteboard based on the virtual change may be projected on top of the physical whiteboard.
G06V 10/22 - Image preprocessing by selection of a specific region containing or referencing a pattern; Locating or processing of specific regions to guide the detection or recognition
An image capture device includes a first housing, a second housing, a first integrated sensor-lens assembly (ISLA), and a second ISLA. The second housing is coupled to the first housing to form an internal compartment. The first ISLA includes a first image sensor coupled to a first lens in fixed alignment. The second ISLA includes a second image sensor coupled to a second lens in fixed alignment. The first ISLA is positively statically connected to the first housing, and the second ISLA is coupled to the first housing indirectly via the first ISLA.
An image capture device includes a processor for wind noise processing. The processor receives signals from microphones. The processor may segment the signals into low frequency bins and high frequency bins. The processor may select a minimum level signal bin for the low frequency bins. The processor may select a minimum level signal bin for the high frequency bins. The processor may generate a composite signal by combining the selected minimum level signal bins for the low frequency bins and the selected minimum level signal bins for the high frequency bins.
H04R 1/40 - Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by combining a number of identical transducers
G10L 25/51 - Speech or voice analysis techniques not restricted to a single one of groups specially adapted for particular use for comparison or discrimination
H04N 23/45 - Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from two or more image sensors being of different type or operating in different modes, e.g. with a CMOS sensor for moving images in combination with a charge-coupled device [CCD] for still images