46 patents in CPC class H04N
Disclosed is an analysis method for detecting and analyzing light from a sample prepared so as to emit light in accordance with an amount of a test substance, the analysis method including taking an image of a storage member configured to store the sample therein; switching a state of a reflector to a state in which light from the sample is reflected toward a light detection unit and detecting light from the sample by the light detection unit; and outputting an analysis result of the sample on the basis of a light amount detected by the light detection unit.
An image processing apparatus includes: an image acquisition unit that acquires a plurality of endoscope images obtained by imaging an observation target at different times with an endoscope; a blood vessel extraction unit that extracts blood vessels of the observation target from the plurality of endoscope images; a blood vessel information calculation unit that calculates a plurality of pieces of blood vessel information for each of the blood vessels extracted from the endoscope images; a blood vessel parameter calculation unit that calculates a blood vessel parameter, which is relevant to the blood vessel extracted from each of the endoscope images, by calculation using the blood vessel information; and a blood vessel change index calculation unit that calculates a blood vessel change index, which indicates a temporal change of the blood vessel, using the blood vessel parameter.
These techniques detect and compare amounts of motion of objects within video footage having sources of apparent motion. In an embodiment, for each frame of a video, a computer subtracts a background from the frame, converts the frame from grayscale to monochrome, and isolates regions of interest (ROIs) in the frame. For each frame, the computer identifies identifiable objects in the ROIs, selects selectable objects from the identifiable objects that have a percentage of motion that is within a particular range, and merges, into merged objects, subsets of the selectable objects that are separated by no more than a particular distance. Based on the merging, metadata is generated for any of: an object of the merged objects, the frame, or the video. The metadata may have a bounding rectangle, a bounding polygon, a frame timestamp, a frame identifier, or an object(s)'s identifier or percentage of motion.
According to an embodiment, a moving object tracking device includes a memory having computer executable components stored therein; and a processor communicatively coupled to the memory. The processor is configured to perform acquiring a plurality of images; generating a plurality of pieces of tracking information indicating information obtained by tracking a moving object included in the images; generating first associated tracking information obtained by first association targeting the plurality of pieces of tracking information having a mutual time difference equal to or smaller than a threshold; generating second associated tracking information obtained by second association targeting the first associated tracking information and the tracking information, not associated by the first association, based on authentication information for identifying the moving object; and outputting the second associated tracking information.
This disclosure describes systems and techniques for generating a disparity map having reduced over-smoothing. To achieve the reduction in over-smoothing that might otherwise occur, a third image is captured in addition to stereo reference and search images. Information from the third image, which may be of the same or different type as the stereo images (e.g., RGB, grey-scale, infra-red (IR)) and which may have the same or different resolution as the stereo images, can help better define the object edges.