Voice-responsive annotation of video generated by an endoscopic camera
First Claim
Patent Images
1. A method comprising:
- receiving a video stream generated by an endoscopic video camera;
receiving and automatically recognizing, by a voice-responsive control system, a spoken utterance of a user while the video stream is being received, wherein the spoken utterance includes a predefined command and additional speech, the voice-responsive control system looks up a non-text annotation corresponding to the additional speech and in response to recognizing the predefined command;
sending, from the voice-responsive control system to an image capture device, a control packet including an indication that the annotation is a non-text visual object, an index of the annotation, and display coordinates for the annotation;
providing, by the image capture device, the video stream and the annotation to a display device for display, such that the annotation is overlaid on a frame of the video stream displayed on the display device at the display coordinates specified by the control packet to point to or outline an anatomical feature; and
associating, by the image capture device, the annotation with the video stream.
1 Assignment
0 Petitions
Accused Products
Abstract
An image capture device in an endoscopic imaging system receives a video stream generated by an endoscopic video camera. In response to automatic recognition of a spoken utterance while the video stream is being received from the endoscopic video camera, the image capture device associates with the video stream an annotation that corresponds to the spoken utterance. The image capture device provides the video stream to a display device for display, such that the annotation can be overlaid on one or more frames of the video stream displayed on the display device.
-
Citations
28 Claims
-
1. A method comprising:
-
receiving a video stream generated by an endoscopic video camera; receiving and automatically recognizing, by a voice-responsive control system, a spoken utterance of a user while the video stream is being received, wherein the spoken utterance includes a predefined command and additional speech, the voice-responsive control system looks up a non-text annotation corresponding to the additional speech and in response to recognizing the predefined command; sending, from the voice-responsive control system to an image capture device, a control packet including an indication that the annotation is a non-text visual object, an index of the annotation, and display coordinates for the annotation; providing, by the image capture device, the video stream and the annotation to a display device for display, such that the annotation is overlaid on a frame of the video stream displayed on the display device at the display coordinates specified by the control packet to point to or outline an anatomical feature; and associating, by the image capture device, the annotation with the video stream. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17)
-
-
18. An apparatus comprising:
-
a voice-responsive control system to receive a video stream generated by an endoscopic video camera; receive and automatically recognize a spoken utterance of a user while the video stream is being received, wherein the spoken utterance includes a predefined command and additional speech, the voice-responsive control system including an annotation dictionary to store a set of annotations, look up, in the annotation dictionary, a non-text annotation corresponding to the additional speech in response to recognizing the predefined command generate a control packet including an indication that the annotation is a non-text visual object, an index of the annotation, and display coordinates for the annotation; and an image capture device to receive the control packet and the video stream from the voice-responsive control system, provide the video stream and the annotation to a display device, such that the annotation is overlaid on a frame of the video stream displayed on the display device at the display coordinates specified by the control packet to point to or outline an anatomical feature, and associate the annotation with at least a portion of the video stream. - View Dependent Claims (19, 20, 21, 22, 23, 24, 25, 26, 27, 28)
-
Specification