Hand sanitation compliance enforcement systems and methods
An electronic processor is operatively connected with at least one video camera, or additionally or alternatively a non-camera-based sensor, and at least one annunciator to process real time video acquired by the at least one video camera in real time to perform a hand hygiene compliance enforcement method for encouraging use of a hand sanitation station.
Hand sanitation compliance enforcement systems and methods
An electronic processor is operatively connected with at least one video camera, or additionally or alternatively a non-camera-based sensor, and at least one annunciator to process real time video acquired by the at least one video camera in real time to perform a hand hygiene compliance enforcement method for encouraging use of a hand sanitation station. The real time video is analyzed to determine an occupancy time interval over which a sanitation zone imaged by the real time video is occupied. Whether a hand sanitation event occurs is determined based on images of a wash station zone extracted from the real time video acquired during the occupancy time interval. The at least one annunciator is controlled to annunciate a hand hygiene compliance status based on the determination of whether a hand sanitation event occurs.
This application claims the benefit of U.S. application Ser. No. 17/067,172 (filed on Oct. 9, 2020) and U.S. Provisional Application No. 62/912,787 (filed on Oct. 9, 2019). The entirety of each of the foregoing applications is incorporated by reference herein.
FIELD OF THE DISCLOSUREThe present disclosure generally relates to hand sanitation compliance enforcement systems and methods, and, more particularly, to hand sanitation and/or hygiene compliance enforcement systems, devices, and/or methods relating to the hand sanitation arts and to related arts such as medical sanitation arts, culinary sanitation arts, and the like.
BACKGROUNDHand sanitation is a critical component in areas such as healthcare, food preparation/handling, child care, restaurants, elderly care, or the like. For example, a hospital ward housing critically ill patients may have a hand sanitation station located outside the entrance door, including a wash basin, sanitizing soap, and/or the like. Doctors, nurses, or other medical personnel are expected to follow a hand hygiene protocol which includes sanitizing their hands at the hand sanitation station prior to entering the hospital ward. In some protocols, gloves may be donned after performing hand sanitation, for example in accordance with a “gown and glove” procedure such as is employed in a clostridium difficule (C-DIFF) protocol. A second hand sanitation station may be located inside the hospital ward proximate to the exit door, and the hand hygiene protocol further calls for medical personnel to sanitize their hands at the second hand sanitation station prior to exiting the hospital ward. Following the prescribed hand hygiene protocol can prevent contact-mediated spread of infectious diseases (e.g., coronavirus disease 2019 aka “COVID-19”) to and/or from critically ill patients. Similar hand hygiene protocols may be established for restaurant workers or others employed in the food preparation/handling industries, for teachers who come into frequent contact with children, for workers at assisted care facilities who come into frequent contact with elderly residents, or so forth.
Unfortunately, compliance with a hand hygiene protocol is imperfect. This can lead to unnecessary spread of infectious diseases (e.g., COVID-19). Noncompliance has been addressed by educational/training programs, and by use of prominent signage to remind personnel to use the hand sanitation station. Education or training, while helpful, does not effectively prevent noncompliance due to forgetfulness or simple laziness. Prominent signage may reduce forgetfulness, but signage can easily become overfamiliar if overused. For example, a hospital may have reminder signs posted at the entrance to every patient ward; over time, hospital personnel come to tune out these ubiquitous signs thereby reducing their effectiveness. Furthermore, prominent signage is not a deterrent to noncompliance due to laziness.
For the foregoing reasons, certain improvements are disclosed herein, as there is a need for systems, devices, and methods for hand hygiene compliance and enforcement.
BRIEF SUMMARYIn accordance with some illustrative embodiments disclosed herein, a hand hygiene compliance enforcement device comprises at least one video camera, at least one annunciator, an electronic processor; and a non-transitory storage medium that stores instructions that are readable and executable by the electronic processor to process real time video acquired by the at least one video camera in real time. The instructions include occupancy detection instructions readable and executable by the electronic processor to analyze the real time video to determine an occupancy time interval over which a sanitation zone imaged by the real time video is occupied. The instructions further include sanitation detection instructions readable and executable by the electronic processor to determine whether a hand sanitation event occurs based on images of a wash station zone extracted from the real time video acquired during the occupancy time period and to control the at least one annunciator to annunciate a hand hygiene compliance status based on the determination of whether a hand sanitation event occurs.
In accordance with some illustrative embodiments disclosed herein, a hand sanitation station includes a hand sanitation station including at least a sanitation fluid dispenser, and a hand hygiene compliance enforcement device as set forth in the immediately preceding paragraph, in which the at least one video camera is arranged such that at least a portion of the hand sanitation station coincides with the wash station zone in the real time video acquired by the at least one video camera.
In accordance with some illustrative embodiments disclosed herein, a non-transitory storage medium stores instructions that are readable and executable by an electronic processor operatively connected with at least one video camera and at least one annunciator to process real time video acquired by the at least one video camera in real time to perform a hand hygiene compliance enforcement method comprising: analyzing the real time video to determine an occupancy time interval over which a sanitation zone imaged by the real time video is occupied; determining whether a hand sanitation event occurs based on images of a wash station zone extracted from the real time video acquired during the occupancy time period; and controlling the at least one annunciator to annunciate a hand hygiene compliance status based on the determination of whether a hand sanitation event occurs.
In accordance with some illustrative embodiments disclosed herein, a hand hygiene compliance enforcement method comprise: using at least one video camera, acquiring real time video of a field of view that includes a sanitation station having a wash station zone; using an electronic processor, determining whether a hand sanitation event occurs based on images of the wash station zone extracted from the real time video; and using an annunciator, annunciating a hand hygiene compliance status based on the determination of whether a hand sanitation event occurs.
As described herein, a hand hygiene compliance enforcement device is disclosed that comprises: at least one video camera; at least one annunciator; an electronic processor; and a non-transitory storage medium storing instructions that are readable and executable by the electronic processor, wherein the instructions, when executed by the electronic processor, cause the electronic processor to: process real time video acquired by the at least one video camera in real time, analyze, by execution of occupancy detection instructions, the real time video to determine an occupancy time interval over which a sanitation zone imaged by the real time video is occupied, determine, by execution of sanitation detection instructions, whether a hand sanitation event occurs based on images of a wash station zone as extracted from the real time video acquired during the occupancy time interval, and control the at least one annunciator to annunciate a hand hygiene compliance status based on the determination of whether a hand sanitation event occurs.
In additional embodiments, as described herein, a non-transitory storage medium is disclosed for storing instructions for performing a hand hygiene compliance enforcement, the instructions being readable and executable by an electronic processor, wherein the instructions, when executed by the electronic processor, cause the electronic processor to: receive real time video acquired by at least one video camera in real time; analyze the real time video to determine an occupancy time interval over which a sanitation zone imaged by the real time video is occupied; determine whether a hand sanitation event occurs based on images of a wash station zone as extracted from the real time video acquired during the occupancy time interval; and control at least one annunciator to annunciate a hand hygiene compliance status based on the determination of whether a hand sanitation event occurs.
In still further embodiments, as described herein, a hand hygiene compliance enforcement method is disclosed that comprises: acquiring, via a at least one video camera, real time video of a field of view that includes a sanitation station having a wash station zone; determining, via an electronic processor, whether a hand sanitation event occurs based on images of the wash station zone as extracted from the real time video; and annunciating, via an annunciator, a compliance status based on the determination of whether a hand sanitation event occurs.
In additional embodiments, as described herein, a hand hygiene compliance enforcement device is disclosed that comprises: at least one non-camera-based sensor; at least one annunciator; an electronic processor; and a non-transitory storage medium storing instructions that are readable and executable by the electronic processor, wherein the instructions, when executed by the electronic processor, cause the electronic processor to: process real time data output of the non-camera-based sensor in real time, analyze, by execution of occupancy detection instructions, the data output to determine an occupancy time interval over which a sanitation zone determined by the data output is occupied, determine, by execution of sanitation detection instructions, whether a hand sanitation event occurs based on the data output corresponding to a wash station zone, the data output acquired during the occupancy time interval, and control the at least one annunciator to annunciate a hand hygiene compliance status based on the hand sanitation event.
As described herein, the embodiments of the disclosure may comprise one or more video cameras and/or, additionally or alternatively, one or more non-camera based sensors. For example, while the disclosure herein refers, in many embodiments, to use of a video camera, it is to be understood that, additionally or alternatively, other motion or video-like devices could be used as a substitute for a video camera. For example, for embodiments using video camera(s), one or more images or image frames (e.g., comprising a video) as captured by the video camera(s) may turned into, or used to create, arrays of digits that represent the image. Additionally, or alternatively, for embodiments using non-camera based sensor(s) the same (or similar) array structure(s) (as created for video camera embodiments) can be emulated by non-camera based sensor(s). Such non-camera based sensor(s) can comprise, by way of non-limiting example, microwave sensor(s), depth sensor(s), tomographic sensor(s), and/or infrared (passive and active) sensor(s), or other such similar or related sensors or technologies. These non-camera based sensor(s) can provide stream(s) of data, or otherwise inputs that emulate the use of a video camera. Just as video camera input, such non-camera based sensor(s) inputs or streams of data can also serve as the base for artificial intelligence based model(s) (e.g., machine learning model(s) and/or CNN model(s)), and related model pattern recognition, as described herein. In this way, non-camera-based sensors can create (or emulate) the same (or similar) array data structure(s) that can then be used to generate tensors for pattern recognition, just as for video camera related embodiments.
In accordance with the above, and with the disclosure herein, present disclosure includes improvements in computer functionality or in improvements to other technologies at least because the claims recite that, e.g., a hand hygiene compliance enforcement device is improved by implementation of machine learning to improve the prediction accuracy of the hand hygiene compliance enforcement device to detect hand sanitation event(s) in real time. That is, the present disclosure describes improvements in the functioning of the computer itself or “any other technology or technical field” because a hygiene compliance enforcement device is better able to predict or detect hand sanitation event(s). This improves over the prior art at least because either no such device(s) existed or were limited to non-machine learning based approaches.
The present disclosure includes applying certain of the features described herein with, or by use of, a particular machine, e.g., a video camera and/or a non-camera based sensor for processing real time video or sensor data acquired by a video camera or non-camera based sensor in real time.
The present disclosure includes effecting a transformation or reduction of a particular article to a different state or thing, e.g., transforming or reducing video images and/or sensor data to actual image or audible output to indicate whether or not a hand sanitation event occurs.
The present disclosure includes specific features other than what is well-understood, routine, conventional activity in the field, or adding unconventional steps that confine the claim to a particular useful application, e.g., hand hygiene compliance enforcement devices and methods, as described herein.
Advantages will become more apparent to those of ordinary skill in the art from the following description of the preferred embodiments which have been shown and described by way of illustration. As will be realized, the present embodiments may be capable of other and different embodiments, and their details are capable of modification in various respects. Accordingly, the drawings and description are to be regarded as illustrative in nature and not as restrictive.
The Figures described below depict various aspects of the system(s), devise(s), and method(s) disclosed therein. It should be understood that each Figure depicts an embodiment of a particular aspect of the disclosed system and methods, and that each of the Figures is intended to accord with a possible embodiment thereof. Further, wherever possible, the following description refers to the reference numerals included in the following Figures, in which features depicted in multiple Figures are designated with consistent reference numerals.
There are shown in the drawings arrangements which are presently discussed, it being understood, however, that the present embodiments are not limited to the precise arrangements and instrumentalities shown, wherein:
The Figures depict preferred embodiments for purposes of illustration only. Alternative embodiments of the systems and methods illustrated herein may be employed without departing from the principles of the invention described herein.
Any quantitative dimensions shown in the drawing are to be understood as non-limiting illustrative examples. Unless otherwise indicated, the drawings are not to scale; if any aspect of the drawings is indicated as being to scale, the illustrated scale is to be understood as non-limiting illustrative example.
DETAILED DESCRIPTIONDisclosed herein are hand hygiene protocol enforcement systems, devices, instructions (as stored on computer readable medium), and methods which improve compliance with a hand hygiene protocol. The disclosed hand hygiene protocol enforcement directly detects compliance or noncompliance, and provides an immediate and active reminder upon detection of noncompliance, for example in the form of a flashing light and/or audible reminder. Unlike static posted signage which is easily tuned out due to its static ubiquity, the flashing light and/or audible reminder is virtually certain to be noticed. Some embodiments of the disclosed hand hygiene protocol enforcement also leverage group psychology by placing the flashing light in a prominent location so that passers-by are likely to notice it and thereby become aware of the noncompliance event; likewise, an audible alarm can be made sufficiently loud to be heard by passers-by. In this way, it is not necessary to specifically identify noncompliant personnel (although identification of personnel as part of the hand hygiene protocol enforcement is contemplated in some variant embodiments) Likewise, the disclosed hand hygiene protocol enforcement can optionally maintain compliance statistics which are anonymous as monitored individuals are not identified (although, again, in some variant embodiments individuals may be identified and the statistics maintained for individuals). A further advantage of the disclosed hand hygiene protocol enforcement is that it is easily retrofitted to existing hand sanitation stations, and can be added to existing or new hand sanitation stations as an after-market modification.
With reference to
With continuing reference to
The video camera 20 is arranged, positioned, or configured such that at least a portion of the hand sanitation station 10 coincides with a wash station zone 40 captured by (i.e. imaged by) real time video acquired by the video camera 20. The wash station zone 40 captured by the video camera 20 (which is only a portion of the full field of view, i.e. FOV, of the video camera 20) thus coincides with the spatial region within which the hand sanitation is expected to be performed by a person using the hand sanitation station 10. In the illustrative example, the wash station zone 40 coincides with the area of the dispenser tip of the hand sanitation fluid dispenser 12, the faucet 14, and the area above the washbasin 16.
The electronic processor 24 and non-transitory storage medium 26 are shown diagrammatically in
As noted, the non-transitory storage medium 26 stores instructions 30, 32, 34 that are readable and executable by the electronic processor
24 to process real time video acquired by the video camera 20 in real time. As diagrammatically indicated in
The annunciator 22 typically includes the aforementioned noncompliance light 22, which is controlled by execution of the sanitation detection instructions 32 to illuminate to indicate detection of noncompliance with the hand hygiene protocol. Typically, the noncompliance light 22 is a red light, although other colors are contemplated herein and may be used. Optionally, the annunciator may include a compliance light 42, e.g. a green light, which is controlled by execution of the sanitation detection instructions 32 to illuminate or to indicate detection of compliance with the hand hygiene protocol (e.g., detection of occurrence of a hand sanitation event). The illustrative lighting annunciators 22, 42 are shown as being placed above the door 18, although they may be placed in some other prominent location where the person entering the door 18 is unlikely to miss them. Additionally or alternatively, the annunciator may be implemented as an audio speaker 44 which is controlled by execution of the sanitation detection instructions 32 to output a verbal reminder to use the hand sanitation station 10 in response to detection of noncompliance with the hand hygiene protocol. Other annunciator devices are additionally or alternatively contemplated, such as a display that displays a textual and/or graphical reminder to use the hand sanitation station 10 in response to detection of noncompliance with the hand hygiene protocol. In lighting or display annunciators, illumination may be accomplished by a flashing illumination (especially in the case of the noncompliance annunciator light 22), or, in the case of a display, showing or displaying the textual and/or graphical reminder as a flashing text and/or flashing graphic.
With reference now to
Additionally, or alternatively, with respect to
In the illustrative example of
Upon detection of motion in the video, process flows to operation 52 where an image is acquired (or, alternatively, the most recent video frame processed in operation 50 is used). At this stage, an image area 54 is considered, which includes the portion of the image depicting the sanitation station 10, the door 18, and closely adjacent area. In an operation 56, image analysis is performed on the image area 54 to confirm that the detected motion is due to a person being in the image area 54. This analysis may, for example, entail computing a difference image for the area 54 comparing the image from operation 52 with earlier frames for which no motion was detected in operation 50, or alternatively computing the difference image respective to a stored reference image of the image area 54. Any suitable image analysis technique for detecting a person in an image can be employed, e.g. connectivity analysis may be performed on the difference image to detect object size and aspect ratio which is compared with expected size and aspect ratios for typical human beings. If no person is detected, flow returns to operation 50 to continue monitoring the video for motion detection.
On the other hand, if the operation 56 confirms a person is in the image area 54, then in an operation 58 it is determined whether a sanitation zone is occupied by the person. The sanitation zone is an area within the area 54 which corresponds to the person being sufficiently proximate to the sanitation station 10 that the person would be expected to be using the sanitation station 10 prior to passing through the door 18. Preferably, the sanitation zone extends from the spatial point where a person might first be close enough to use the sanitation station 10 to the spatial point where the person has fully passed the sanitation station 10 such that it becomes apparent that the person has bypassed the sanitation station 10. Various approaches can be used to determine the occupancy. For example, for successive frames of the real time video, an image of the sanitation zone can be extracted from the frame and compared with a reference image of the sanitation zone to detect whether the extracted image of the sanitation zone indicates occupation of the sanitation zone. Additionally or alternatively, for successive frames of the real time video, an image of the sanitation zone can be extracted from the frame. The images of the sanitation zone extracted from the successive frames of the real time video are then compared to detect frame-to-frame changes that indicate occupation of the sanitation zone. Again, each of these are merely illustrative examples.
With conditioned reference to
Although not explicitly depicted in
With continuing reference to
Machine learning model(s), such as those of a machine learning model, CNN, or other AI model as described herein, may be created and trained based upon example (e.g., “training data,”) inputs or data (which may be termed “features” and “labels”) in order to make valid and reliable predictions for new inputs, such as testing level or production level data or inputs. In supervised machine learning, a machine learning program operating on a server, computing device, or otherwise processor(s), may be provided with example inputs (e.g., “features”) and their associated, or observed, outputs (e.g., “labels”) in order for the machine learning program or algorithm to determine or discover rules, relationships, or otherwise machine learning “models” that map such inputs (e.g., “features”) to the outputs (e.g., labels), for example, by determining and/or assigning weights or other metrics to the model across its various feature categories. Such rules, relationships, or otherwise models may then be provided subsequent inputs in order for the model, executing on the server, computing device, or otherwise processor(s), to predict, based on the discovered rules, relationships, or model, an expected output.
In unsupervised machine learning, the server, computing device, or otherwise processor(s), may be required to find its own structure in unlabeled example inputs, where, for example multiple training iterations are executed by the server, computing device, or otherwise processor(s) to train multiple generations of models until a satisfactory model, e.g., a model that provides sufficient prediction accuracy when given test level or production level data or inputs, is generated. The disclosures herein may use one or both of such supervised or unsupervised machine learning techniques.
With reference to
Alternatively, the training may use training images or sensor data of a standard wash station zone that depict a hand sanitation event and training images of the standard wash station zone that do not depict a hand sanitation event. To generate training images or sensor data, video is taken, or sensor data is collected, of various people passing through the sanitation zone of a standard sanitation station (preferably of the same manufacture and model as the sanitation station 10 at which the compliance enforcement device is to be deployed) who do, or do not, use the standard sanitation station. This approach advantageously allows for the pre-trained CNN (or other ML classifier) 60 to be shipped with the compliance enforcement device and eliminates the need to perform on-site training of the CNN.
Rather than using the CNN 60, another type of ML classifier could be used, such as another type of artificial neural network (ANN), or a support vector machine (SVM) classifier, or another architecture
such as VGG, ResNET, RCNN, MaskRCNN, MobileNET, Inception family or so forth. In another approach, the CNN 60 is trained to employ Human Pose estimation, in which the CNN 60 is
trained on key points that align to human body parts such as the right elbow, left knee, and so forth. Using a CNN for Human Pose estimation advantageously allows for deriving when a hand sanitation action would
occur. In tests, this approach was significantly more computationally heavy and less accurate, making real time inference difficult. Additionally or alternatively, a local node of cloud computation can be used,
which may allow the systems and/or methods of
With continuing reference to
In the illustrative example of
In a variant annunciation approach, the audible initial reminder via the loud speaker 44 (if provided) may initially be provided as the person enters the sanitation zone, but the noncompliance light 22 is not initially activated. Thereafter, as the person passes through the sanitation zone no annunciation is provided (e.g., neither light 22, 42 is lit, nor is any audible annunciation output). If at some point the person uses the sanitation station 10 then this will be detected at operation 62 causing the (e.g. green) compliance light 42 (if provided) is turned on. On the other hand, if the person completes passage through the sanitation zone without using the sanitation station 10 then the outcome operation 70 illuminates the non-compliance light 22 (for the first time, in this embodiment) to remind the person of his or her noncompliance, and/or issues an audible reminder via the loud speaker 44. This approach provides a more delayed response, but has certain advantages. For example, in this embodiment the red light 22 only illuminates if the person passes completely through the sanitation zone without using the sanitation station 10. In other words, the red light 22 only illuminates if the person has failed to use the sanitation station 10. This embodiment advantageous leverages social pressure to encourage enforcement with the hand hygiene protocol, because if the red light 22 illuminates then any person in the area who sees it will know the person approaching the door 18 has failed to use the sanitation station 10.
Setup of the hand hygiene compliance enforcement device may be done at some time after the sanitation station 10 is installed, and proceeds as follows. The video camera 20 and annunciator(s)
22, 42, 44 are installed, and the electronic processor 24 and storage 26 are installed and operatively connected with the video
camera 20 and annunciator(s) 22, 42, 44. These connections may be by way of physical wires or cables, or by wireless connections (e.g., Bluetooth™,
WiFi, et cetera). The software 30, 32, 34 can be installed, or, can come preinstalled on the storage medium 26. Next, the area
54 (see
In various embodiments, the annunciator(s) 22, 42, 44, and/or a display or touchscreen display may be used for sanitation “information sharing.” Information sharing comprises sharing information within a proximity to a hand hygiene compliance enforcement device, or portion thereof. The information may be shared, e.g., via a touchscreen display or display viewable by people) in the vicinity of the hygiene compliance enforcement device, or portion thereof. Such public display would invoke “information sharing” to encourage individuals, who are not bound (e.g. legally bound) by compliance, such as hospital visitors and others, to nonetheless engage in positive hand sanitation behaviors. For example, unlike, doctors, nurses, and hospital staff, or restaurant workers or lab workers, individuals not required to comply with hand sanitation regulations may fail to comply with hand sanitation, without social pressure as caused by an “information sharing” annunciator, such as a display screen showing a given individual's hand sanitation compliance. By centralizing, or making available, the dissemination of simple information at or near a hand hygiene compliance enforcement device, or portion thereof, information sharing on display devices (or other annunciator(s) as described herein) may cause persons in public spaces that incorporate these device(s) to comply with sanitation regulations, thereby increasing hand sanitation compliance overall. For example, in an embodiment, a touchscreen display may be used to share information about a given hospital's hand sanitation compliance, e.g., time, rating, and positive reports regarding hand sanitation that the hospital would could then share with customers, patients, etc. By sharing such information on a display or screen, near a hand hygiene compliance enforcement device, or portion thereof, such information would further encourage positive hand sanitation protocol(s). As a further example, a restaurant may use a screen to share information about sanitation compliance, including the restaurant's use of menus, QR codes, or other touchless payment types that demonstrate restaurant compliance with certain disease protocols (e.g., COVID-19). As a further example, and in general, public spaces, such as museums, co-working spaces, or locations where many people gather, could have installed an information sharing system, as described herein, to not only share compliance “guidelines,” but also to determine who is (and how many people are) are complying, whether it be outside or inside of legal compliance. Such embodiments provide the advantage of increasing compliance because, as demonstrated in these examples, such embodiments may lead to reduced “sanitation leakage,” especially in places where a lack of sanitation could occur because of a typical lack of tracking or monitoring. Accordingly, use of a hygiene compliance enforcement device, or portion thereof, in such locations may cause increased sanitation, and related statistics, and better metrics overall.
In various embodiments, information sharing data or information may be stored in statistics database 80, or in other computer memorie(s) associated with electronic processor such as a non-transitory storage medium 26. The information sharing data or information may be reported as described herein, via reporting instructions 34 and/or compliance report generator 82. In further embodiments, a hand hygiene compliance enforcement device may be configured to control information sharing through an application (app), which may comprise one or more computing instructions stored in non-transitory storage medium 26, and configured for execution on electronic processor 24. The application may be a native app, such as an APPLE iOS app or a GOOGLE ANDROID app. The information sharing app could comprise an information sharing interface as implemented or displayed on a mobile device (e.g., smart phone or tablet) or could comprise a webpage of a website. The information sharing interface could provide data or statistics for information sharing, as described herein, to a display screen, e.g., via reporting instructions 34 and/or compliance report generator 82. In this way, such information sharing data or statistics may be provided to a hygiene compliance enforcement device or, additionally or alternatively, to a group of devices (e.g., mobile devices) via one or more corresponding information sharing interface(s). By using the information sharing interface(s), a customer or any person that wants to view or share information regarding hand sanitation compliance could access an information sharing interface, e.g., by downloading an app, uploading an image, uploading a video, or sharing text content, to provide and/or monitor hand sanitation compliance of a particular zone or location, even a location remote to the hygiene compliance enforcement device or the user.
In some embodiments, information sharing and display, e.g., via information sharing interface(s) as described herein, may comprise displaying notifications, marketing, advertisements, or the like. In such embodiments, “marketing verification inference analysis” may be provided as, or with, hand sanitation messages or other notifications as described herein. In particular, tracking marketing inferences in a physical marketing environment is typically difficult for marketers. Through hand sanitation compliance, as described herein, however, tracking inference and physical presence of individuals could be implemented by use of hand hygiene compliance enforcement devices as described herein. For example, display of advertisement(s) and related tracking as to one or more marketing inference(s) across one or more users may be stored on statistics database 80, or in other computer memorie(s) associated with electronic processor such as a non-transitory storage medium 26. The display and/or tracking may be implemented though the hand sanitation techniques as described herein. For example, the annunciators and related imaging analysis, such as provided via a display and/or touchscreen display, which may be in a proximity to a hand sanitation zone and/or device as described herein, may be used for such tracking.
For example,
These advertisements can be controlled through a hand hygiene marketing platform, which may be implemented on the hand hygiene compliance enforcement device described herein. The platform can comprise a sanitation marketing application that connects marketers through the hand hygiene compliance enforcement device. The hand hygiene marketing platform may be configured to control marking through a marketing application (app), which may comprise one or more computing instructions stored in non-transitory storage medium 26, and configured for execution on electronic processor 24. The hand hygiene marketing platform could comprise a marketing interface as implemented or displayed on a mobile device (e.g., smart phone or tablet) or could comprise a webpage of a website, or native app as described herein. The marketing interface provides data or statistics for providing advertisement, as described herein, to a display screen, e.g., via reporting instructions 34 and/or compliance report generator 82. In this way, such advertisements may be provided to a hygiene compliance enforcement device or, additionally or alternatively, to a group of devices (e.g., mobile device) via one or more corresponding display or screen interface(s).
In various contemplated approaches, a CNN or other ML architecture can be used to define the wash station and sanitation zones, e.g. the CNN 60 could be another type of ANN, or an SVM classifier, or could be trained to employ Human Pose estimation, or so forth. Finally, as previously noted, the CNN (or other ML classifier) 60 is pretrained to generate the confidence value using training images of the wash station zone or a standard wash station zone that depict a hand sanitation event and training images of the wash station zone or the standard wash station zone that do not depict a hand sanitation event. If training images of a standard wash station zone are used, then the pretrained CNN 60 comes preloaded with the sanitation detection software module 32. On the other hand, if training images of the wash station zone 40 of the sanitation station 10 are to be used in the training, then the installer collects the requisite training data. Preferably, the training data are collected using a number of different people capturing the expected range of human variation (height, weight, skin tone, et cetera) commensurate with the full range of people expected to be using the sanitation station 10. In this latter case, the loaded software includes the CNN training software so that the CNN 60 can be trained on the just-collected training images. Finally, the installer tests the compliance enforcement system using several test runs (again, preferably with people of different heights, weights, skin tones, et cetera) to verify accuracy of the compliance enforcement system.
With continuing reference to
With continuing reference to
With reference now to
The region around a wash station is typically busy with nurses and doctors rushing from room to room and ward to ward. When looking at a full video frame (e.g., for 720 p video a frame has a height of 720 pixels and a width of 1,280 pixels) there are many different patterns from the various moving persons. If the CNN 60 receive full frames (e.g. 720×1280 pixel frames of 720 p video) then detecting a difference between a full frame capturing a hand sanitation event versus a full frame that does not capture a hand sanitation event, the CNN 60 will see many patterns and correlations that are not necessarily relevant to the difference outcome. Incorrect correlation of a non-relevant pattern to a relevant outcome pattern can lead to bias and inaccuracies. By cropping the image processed by the CNN 60 to the area around the hand sanitation station, the CNN is fed the essential information and patterns relevant to detecting a hand sanitation event. However, if taken too far this can also lead to an inaccurate classifier. When the size of the crop is reduced, relevant patterns may be removed. The relevance of these removed patterns may not be apparent to the designer of the classifier, but would be useful input for improving the accuracy of the classifier. By way of a specific example, only looking at the sanitation zone may cut out a window or other reflective surface that the designer might not recognize to be relevant for detecting a hand sanitation event, but that might in fact improve the classifier accuracy. Hence, the size of the wash station zone 40 should preferably include all important relevant patterns, while reducing or eliminating non-relevant patterns. In addition to accuracy, computing on less pixels increases overall system speed.
In addition to design of the pixel dimensions of the wash station zone 40, another way to improve classifier accuracy is to provide the classifier 60 with an image having clean patterns and low noise to the extent possible. This can be achieved by smoothing without losing gradients containing relevant information in the image. Smoothing is a computer vision mechanism based on kernel mathematics to reduce or eliminate high-frequency, outlier heavy aspects of an image. An approximating function should be used that captures important patterns in the data, while leaving out noise or other fine-scale structures that might otherwise be incorrectly matched to patterns indicative of a hand sanitation event. The sample array values of the image are modified so the individual points are reduced and the points that are lower than the adjacent points are increased leading to a smoother signal and less variation. Therefore, a sample series of data points are taken from the image, and a filter or kernel of various sizes and aggregation methods is applied and repeatedly iterated through the image to output a smoothed image. By way of a non-limiting illustrative example, a linear filter may be used for smoothing, which determines the output pixels by a simple weighted value from the surrounding pixels of various size. This is represented, for example, by the below Equation (1):
As shown in Equation (1), the output pixels O(r,g,b) in red, green, and blue, respectively, is equal to the sum of the kernel h(k,l, h) multiplied by the weighted sum of the input pixels f(r+k, g+l, b+k), as expressed in Equation (1). The kernel is, in this example, a matrix with a fixed size 15×15, which looks at the 15 pixels to consider in Equation (1). Therefore, the kernel is just a coefficient to the filter or the “window size” and the filter is a sliding “window” that goes across portions of the image to output different sized and averaged array tensors or “smoother image.” In other words, this image transformation is applied to each pixel in the image to achieve a different state. The goal of this operation is to smooth the image, which can reduce the noise. However, the smoothing can also smooth away the edges and important information and patterns that are relevant for detecting a hand sanitation event. Hence, over-smoothing can lead to an inaccurate classifier and reduced real time accuracy. To address this, the smoothing operation may optionally incorporate an Image Gradient to smooth away noise while retaining the relevant patterns. This problem is solved by looking at the rate of change in the pixels to determine important edges and features. By grouping pixels that are close to one another in value and close to one another in spatial location and leveraging Laplacian, Robers or Prewitt operations these “edges” can be found. In other words, these operations are designed to automatically limit the smoothing at “edges” in images by highly polarized pixel values. Therefore, the gradients build on the smoothing operation to be defined as Bilateral filtering, as expressed, for example, by Equation (2):
As shown in Equation (2) above, I is the original input image, x is the coordinates of the current pixel, Ω is the “isolator window” centered in x, Fr is the range kernel for smoothing differences in intensities, and Gs is the spatial kernel for smoothing difference in coordinates. So by utilizing the rate of change of the two different smoothing operations “un-important” pixels can be smoothed while bringing out “important” pixels. This leads to a pixel window that is optimized for the most relevant patterns to show. This allows a greater chance for the classifier training to attach correct correlation to outcomes, while still allowing for the unpredictable correlations to be present. Therefore these operations affect the training and system side positively. First off, training on higher accurate data and less physical information increases training accuracy, iteration speed and overall computation efficiency. On the inference side cropping, smoothing and gradients allow for less pixels to be consumed which allows for real-time monitoring to be possible, while increasing inference accuracy. The described bilateral filtering is merely an illustrative example, and other types of filtering could be employed to facilitate locating the sanitation zone, for example edge-detection filtering to enhance the strongest edges in the image.
The same down-sampling operations that encompass the smoothing and image gradients can be up-sampled. This means that for a blurry or smaller pixel window the image processing can attempt to make a sharper “better” image. This is done by finding the inverse of the functions set forth above. Therefore, the sanitation zone could be up-sampled or down-sampled to be any pixel range, suitably chosen to optimize accuracy of the classifier by avoiding losing too much relevant information or adding too much noise.
As noted previously, the wash station zone 40 should coincide with the spatial region within which the hand sanitation is expected to be performed by a person using the hand sanitation station 10. During installation, the video camera 20 is positioned to capture the sanitation zone, and the installer delineates the area of a frame that depicts the wash station zone 40; accordingly, the extraction of the image of the wash station zone at operation 102 merely entails extracting that pre-delineated area of the frame. The image may optionally entail further preprocessing to condition the image of the wash station zone for input to the CNN 60. For example, in the illustrative embodiment at an operation 103 the extracted image of the wash station zone is converted to RGB (i.e. red-green-blue) 3×3 channel arrays (tensor). More generally, the image of the wash station zone extracted from the frame 100 may be converted to a tensor of channel arrays wherein each channel array represents a color channel of the image of the wash station zone. The color channels may be RGB color channels as in the illustrative example, or may be color channels of another color space such as LAB, YCrCb, HSV, or so forth. In other embodiments, it is contemplated to replace (or augment) the RGB, YCrCb, HSV, or other color video feed with depth information. A depth camera could be used or a regular camera could be used utilizing a depth map function. The depth of a pixel could be found using disparity=x−x=Bf/Z. Utilizing this, the depth of all pixels in an image can be derived. However, it is expected that employing depth information would increase the computational load and may lead to an image with less distinctive sanitation attributes. Other types of imaging, such as thermal imaging could also be used. In an operation 104, the CNN 60 is applied to the extracted (and optionally preprocessed) image of the wash station zone to generate a confidence value 106 indicative of likelihood that the image of the wash station zone depicts a hand sanitation event. In this regard, it should be noted that the training images of the wash station zone 40 or of a standard wash station zone used to (pre-)train the CNN (or other ML classifier) 60 should have been processed using the same extraction/preprocessing 102. For example, if the preprocessing 102 includes converting to RGB 3×3 channel arrays prior to input to the CNN 60 at operation 104, then the training images should also have been converted to RGB 3×3 channel arrays prior to use in training the CNN.
At an operations 107, 108, the confidence 106 is assessed to determine whether the frame depicts an occurrence of a hand sanitation event. For example, if the confidence
106 is higher than some threshold (e.g., an x-value), then it is determined that a hand sanitation event occurred 64, and the process of
In the above Equation (3), ϕ is a mapping specified as ϕ:n→ which has weights defined as w=(ω1, . . . , ωn)T where the weights or confidence readings, ω1, . . . , ωn, are between 0-1 and the sum of the weights is equal to one, i.e. wi∈[0,1] for all i=1, . . . ,n, and Σi=1nwi=1. Using Equation (3), an aggregation method is formed that helps connect the linguistic outputs (here hand sanitation and no hand sanitation) with unsure values and weights. This is merely one illustrative aggregation method, and other simpler or more complex aggregation methods can be utilized, or other approaches as already mentioned.
On the other hand, if the frame is not sufficient to determine that a hand sanitation event occurred (i.e., if the confidence does not meet the threshold x-value), then in an operation 110 the result
is aggregated with confidence values for earlier frames. At an operation 112, it is determined whether the person is still in the sanitation zone (that is, it is determined whether the occupancy
time interval is continuing or has ended). If the person is still in the sanitation zone then a next frame 114 of the video (which may be the immediately next frame in the video stream providing
a 30 Hz time resolution at 30 fps, or the fifth next frame providing a 6 Hz resolution at 30 fps, or so forth) is obtained and processed by the operations 102, 104,
108, 110 as described. This continues until the occupancy time interval ends, at which time the operation 112 will determine the person is no longer in the sanitation
zone. At this point, in an operation 116 the aggregation of confidences from the operation 110 is examined to determine whether it can be concluded that no hand sanitation event
occurred over the occupancy time interval. This can employ various statistical approaches. In one approach, the largest confidence from any of the processed frames is compared with a threshold y-value, and if it
is below this threshold y-value then it is determined that no hand sanitation event occurred 66, and the process of
In the operations 108, 116, the respective thresholds x-value and y-value may be suitably optimized during the training of the CNN (or other ML classifier) 60, e.g. being an additional two parameters to be optimized in addition to the activation weights and/or other parameters of the CNN 60.
The illustrative embodiment of the operation 62 presented with reference to
The illustrative embodiments describe detection of a hand sanitation event via analysis of extracted images of the wash station. In the case of a “gown and glove” procedure such as is employed in a C-DIFF protocol,
it is additionally contemplated to employ similar analysis of an image of a gowning zone (e.g., proximate to a closet containing gowns to be put on during the gowning part of the C-DIFF protocol) to detect a gowning
event, so as to monitor compliance with both the gowning and gloving aspects of the C-DIFF or similar gown-and-glove procedure. In this case, the CNN or other ML classifier is trained using training videos of persons
gowning (positive examples) or failing to gown (negative examples) in a manner analogous to the training described for the illustrative CNN 60. The electronic processor
24 is used to determine whether a gowning event occurs based on images of a gowning zone extracted from the real time video, and the compliance status (variant of 64 or
66 of
The illustrative examples employ a single video camera 20 as shown in
The preferred embodiments have been illustrated and described. Obviously, modifications and alterations will occur to others upon reading and understanding the preceding detailed description. It is intended that the invention be construed as including all such modifications and alterations insofar as they come within the scope of the appended claims or the equivalents thereof.
Aspects of the Present Disclosure
The following aspects of the disclosure are exemplary only and not intended to limit the scope of the disclosure.
1. A hand hygiene compliance enforcement device comprising: at least one video camera; at least one annunciator; an electronic processor; and a non-transitory storage medium storing instructions that are readable and executable by the electronic processor to process real time video acquired by the at least one video camera in real time, the instructions including: occupancy detection instructions readable and executable by the electronic processor to analyze the real time video to determine an occupancy time interval over which a sanitation zone imaged by the real time video is occupied; and sanitation detection instructions readable and executable by the electronic processor to determine whether a hand sanitation event occurs based on images of a wash station zone extracted from the real time video acquired during the occupancy time period and to control the at least one annunciator to annunciate a hand hygiene compliance status based on the determination of whether a hand sanitation event occurs.
2. The hand hygiene compliance enforcement device of aspect 1, wherein the sanitation detection instructions are readable and executable by the electronic processor to determine whether a hand sanitation event occurs by operations including: for each frame of successive frames of the real time video acquired during the occupancy time period, applying a machine learning (ML) classifier to the image of the wash station zone extracted from the frame to generate a confidence value indicative of likelihood that the image of the wash station zone depicts a hand sanitation event; and determining whether a hand sanitation event occurs in the wash station zone based on the confidence values for the successive frames, wherein the ML classifier is pre-trained to generate the confidence value using training images of the wash station zone or a standard wash station zone that depict a hand sanitation event and training images of the wash station zone or the standard wash station zone that do not depict a hand sanitation event.
3. The hand hygiene compliance enforcement device of aspect 2, wherein the applying of the ML classifier includes: converting the image of the wash station zone extracted from the frame to a tensor of channel arrays wherein each channel array represents a color channel of the image of the wash station zone; and applying the ML classifier to the tensor of channel arrays.
4. The hand hygiene compliance enforcement device of any one of aspects 1-3, wherein the images of the wash station zone extracted from the real time video acquired during the occupancy time period have pixel dimensions L×W where L is 100 pixels or less and W is 100 pixels or less.
5. The hand hygiene compliance enforcement device of aspect 4, wherein the extracting of the images of the wash station zone from the real time video acquired during the occupancy time period includes pixel resampling, the pixel dimensions L×W being after the pixel resampling.
6. The hand hygiene compliance enforcement device of any one of aspects 1-5, wherein the sanitation detection instructions are readable and executable by the electronic processor to determine whether a hand sanitation event occurs in the wash station zone by operations including: detecting a color of a hand sanitation fluid in the images of the wash station zone extracted from the real time video acquired during the occupancy time period.
7. The hand hygiene compliance enforcement device of any one of aspects 1-6, wherein the occupancy detection instructions are readable and executable by the electronic processor to analyze the real time video to determine the occupancy time interval by operations including at least one of: for successive frames of the real time video, extracting an image of the sanitation zone from the frame and comparing the extracted image of the sanitation zone with a reference image of the sanitation zone to detect whether the extracted image of the sanitation zone indicates occupation of the sanitation zone; and/or for successive frames of the real time video, extracting an image of the sanitation zone from the frame and comparing the images of the sanitation zone extracted from the successive frames of the real time video to detect frame-to-frame changes that indicate occupation of the sanitation zone.
8. The hand hygiene compliance enforcement device of any one of aspects 1-7, wherein the at least one annunciator comprises a noncompliance light and the sanitation detection instructions are readable and executable by the electronic processor to annunciate the hand hygiene compliance status by at least one of: a) controlling the noncompliance light to illuminate starting at the beginning of the occupancy time interval and ending at the earlier of (i) a stop time respective to the end of the occupancy time interval or (ii) a time when it is determined that a hand sanitation event occurs; b) controlling the noncompliance light to illuminate starting at the end of the occupancy time interval if it is determined that no hand sanitation event occurred during the occupancy time interval.
9. The hand hygiene compliance enforcement device of aspect 8, wherein the at least one annunciator further comprises a compliance light and the sanitation detection instructions are readable and executable by the electronic processor to further annunciate the hand hygiene compliance status by at least one of: a) controlling the compliance light to illuminate starting at a time in the occupancy time interval when it is determined that a hand sanitation event occurs; b)
controlling the compliance light to illuminate starting at the end of the occupancy time interval if it is determined that a hand sanitation event occurred during the occupancy time interval.
10. The hand hygiene compliance enforcement device of any one of aspects 8-9, wherein the noncompliance light is illuminated by a flashing illumination.
11. The hand hygiene compliance enforcement device of any one of aspects 1-10, wherein the at least one annunciator includes at least one of: an at least one light and/or an audio speaker and/or a display.
12. The hand hygiene compliance enforcement device of any one of aspects 1-11, wherein the instructions stored on the non-transitory storage medium that are readable and executable by the electronic processor to process the real time video acquired by the at least one video camera do not include instructions readable and executable by the electronic processor to identify a person in the sanitation zone.
13. The hand hygiene compliance enforcement device of aspect 12, wherein the instructions further include: reporting instructions readable and executable by the electronic processor to accumulate statistics for a plurality of successive occupancy time periods on corresponding determinations of whether a hand sanitation event occurs and generate a compliance report on the accumulated statistics.
14. The hand hygiene compliance enforcement device of any one of aspects 1-13, wherein the instructions further include: reporting instructions readable and executable by the electronic processor to identify a person in the sanitation zone during the occupancy time period using a radio frequency identification (RFID) reader that detects RFID badges or facial recognition applied to the real time video; and accumulate statistics for a plurality of successive occupancy time periods on corresponding determinations of whether a hand sanitation event occurs and generate a compliance report on the accumulated statistics including identification of persons identified in the sanitation zone using the facial recognition during occupancy time periods for which the corresponding determination of whether a hand sanitation event occurs is that no hand sanitation event occurred.
15. The hand hygiene compliance enforcement device of any one of aspects 1-14, wherein the at least one video camera consists of a single video camera and the wash station zone is contained in the sanitation zone imaged by the real time video.
16. A hand sanitation station including: a hand sanitation station including at least a sanitation fluid dispenser; and a hand hygiene compliance enforcement device as set forth in any one of aspects 1-15, wherein the at least one video camera is arranged such that at least a portion of the hand sanitation station coincides with the wash station zone in the real time video acquired by the at least one video camera.
17. A non-transitory storage medium storing instructions that are readable and executable by an electronic processor operatively connected with at least one video camera and at least one annunciator to process real time video acquired by the at least one video camera in real time to perform a hand hygiene compliance enforcement method comprising: analyzing the real time video to determine an occupancy time interval over which a sanitation zone imaged by the real time video is occupied; determining whether a hand sanitation event occurs based on images of a wash station zone extracted from the real time video acquired during the occupancy time period; and controlling the at least one annunciator to annunciate a hand hygiene compliance status based on the determination of whether a hand sanitation event occurs.
18. The non-transitory storage medium of aspect 17, wherein the determining of whether a hand sanitation event occurs comprises: applying a Convolutional Neural Network (CNN) to the images of the wash station zone extracted from successive frames of the real time video acquired during the occupancy time period to generate per-frame confidence values; and determining whether a hand sanitation event occurs based on the per-frame confidence values for the successive frames, wherein the CNN is pre-trained to generate confidence values indicating confidence that input images depict a hand sanitation event using training images of the wash station zone or a standard wash station zone that depict a hand sanitation event and training images of the wash station zone or the standard wash station zone that do not depict a hand sanitation event.
19. The non-transitory storage medium of aspect 18, wherein the applying of the CNN includes: converting the image of the wash station zone extracted from the frame to a tensor of channel arrays wherein each channel array represents a principal color channel of the image of the wash station zone; and applying the CNN to the tensor of channel arrays.
20. The non-transitory storage medium of any one of aspects 17-19, wherein the images of the wash station zone extracted from the real time video acquired during the occupancy time period have pixel dimensions L×W where L is 100 pixels or less and W is 100 pixels or less.
21. The non-transitory storage medium of aspect 20, wherein the extracting of the images of the wash station zone from the real time video acquired during the occupancy time period includes pixel resampling, the pixel dimensions L×W being after the pixel resampling.
22. The non-transitory storage medium of any one of aspects 16-21, wherein the hand hygiene compliance enforcement method further comprises: accumulating statistics for a plurality of successive occupancy time periods on corresponding determinations of whether a hand sanitation event occurs; and generating a compliance report on the accumulated statistics.
23. A hand hygiene compliance enforcement method comprising: using at least one video camera, acquiring real time video of a field of view that includes a sanitation station having a wash station zone; using an electronic processor, determining whether a hand sanitation event occurs based on images of the wash station zone extracted from the real time video; and using an annunciator, annunciating a compliance status based on the determination of whether a hand sanitation event occurs.
24. The hand hygiene compliance enforcement method of aspect 23, wherein the determining of whether a hand sanitation event occurs comprises: applying a machine learning (ML) classifier to the images of the wash station zone extracted from successive frames of the real time video to generate per-frame confidence values; and determining whether a hand sanitation event occurs based on the per-frame confidence values for the successive frames, wherein the ML classifier is pre-trained using training images of the wash station zone or a standard wash station zone that depict a hand sanitation event and training images of the wash station zone or the standard wash station zone that do not depict a hand sanitation event.
25. The hand hygiene compliance enforcement method of aspect 24, wherein the applying of the ML classifier includes: converting the image of the wash station zone extracted from the frame to a tensor of channel arrays wherein each channel array represents a principal color channel of the image of the wash station zone; and applying the ML classifier to the tensor of channel arrays.
26. The hand hygiene compliance enforcement method of any one of aspects 23-25 further comprising: analyzing the real time video to determine an occupancy time interval over which a sanitation zone imaged by the real time video is occupied; wherein the determining of whether a hand sanitation event occurs is based on images of the wash station zone extracted from only the real time video acquired during the occupancy time interval.
27. The hand hygiene compliance enforcement method of any one of aspects 23-26 further comprising: using the electronic processor, determining whether a gowning event occurs based on images of a gowning zone extracted from the real time video, wherein the compliance status is further based on the determination of whether a gowning event occurs.
28. The hand hygiene compliance enforcement method of aspect 27, wherein the determining of whether a gowning event occurs comprises: applying a machine learning (ML) classifier to the images of the gowning zone extracted from successive frames of the real time video to generate per-frame gowning event confidence values; and determining whether a gowning event occurs based on the per-frame gowning event confidence values for the successive frames, wherein the ML classifier for generating the per-frame gowning event confidence values is pre-trained using training images of the gowning zone or a standard gowning zone that depict a gowning event and training images of the gowning zone or the standard gowning zone that do not depict a gowning event.
Additional Aspects of the Present Disclosure
The following additional aspects of the disclosure are exemplary only and not intended to limit the scope of the disclosure.
1. A hand hygiene compliance enforcement device comprising: at least one video camera; at least one annunciator; an electronic processor; and a non-transitory storage medium storing instructions that are readable and executable by the electronic processor, wherein the instructions, when executed by the electronic processor, cause the electronic processor to: process real time video acquired by the at least one video camera in real time, analyze, by execution of occupancy detection instructions, the real time video to determine an occupancy time interval over which a sanitation zone imaged by the real time video is occupied, determine, by execution of sanitation detection instructions, whether a hand sanitation event occurs based on images of a wash station zone as extracted from the real time video acquired during the occupancy time interval, and control the at least one annunciator to annunciate a hand hygiene compliance status based on the determination of whether a hand sanitation event occurs.
2. The hand hygiene compliance enforcement device of aspect 1, wherein the electronic processor is configured to execute the sanitation detection instructions to cause the electronic processor to: apply, for each respective frame of successive frames of the real time video acquired during the occupancy time interval, a machine learning (ML) classifier to one or more images of the wash station zone as extracted from the respective frame to generate a confidence value indicative of likelihood that the one or more images of the wash station zone depicts a hand sanitation event, and determine whether a hand sanitation event occurs in the wash station zone based on the confidence values for the successive frames, wherein the ML classifier is pre-trained to generate the confidence value using (1) training images of the wash station zone or a standard wash station zone that depicts a hand sanitation event, and (2) training images of the wash station zone or the standard wash station zone that do not depict a hand sanitation event.
3. The hand hygiene compliance enforcement device of aspect 2, wherein application of the ML classifier comprises: converting an image of the wash station zone as extracted from a frame of the successive frames to a tensor of channel arrays, wherein each channel array represents a color channel of the image of the wash station zone; and applying the ML classifier to the tensor of channel arrays.
4. The hand hygiene compliance enforcement device as in any one of aspects 1-3, wherein the images of the wash station zone, as extracted from the real time video acquired during the occupancy time interval, have pixel dimensions length (L) by width (W) where L is 100 pixels or less and W is 100 pixels or less, wherein extraction of the images of the wash station zone from the real time video acquired during the occupancy time interval comprises pixel resampling, the pixel dimensions L by W being determined after the pixel resampling.
5. The hand hygiene compliance enforcement device as in any one of aspects 1-4, wherein the electronic processor is configured to execute the sanitation detection instructions to cause the electronic processor to: determine whether a hand sanitation event occurs in the wash station zone by detecting a color of a hand sanitation fluid in the images of the wash station zone as extracted from the real time video acquired during the occupancy time interval.
6. The hand hygiene compliance enforcement device as in any one of aspects 1-5, wherein the electronic processor is configured to execute the occupancy detection instructions to cause the electronic processor to perform at least one of: a) extracting, for one or more successive frames of the real time video, an extracted image of the sanitation zone from a frame of the one or more successive frames, and comparing the extracted image of the sanitation zone with a reference image of the sanitation zone to detect whether the extracted image of the sanitation zone indicates occupation of the sanitation zone; or b) extracting, for two or more successive frames of the real time video, an extracted image of the sanitation zone from a respective frame of the two or more successive frames, and comparing a first image of the sanitation zone as extracted from a first respective frame of the two or more successive frames with a second image of the sanitation zone as extracted from a second respective frame of the two or more successive frames to detect frame-to-frame changes that indicate occupation of the sanitation zone.
7. The hand hygiene compliance enforcement device as in any one of aspects 1-6, wherein the at least one annunciator comprises a noncompliance light, and wherein the electronic processor is configured to execute the sanitation detection instructions to cause the electronic processor to perform at least one of: a) controlling the noncompliance light to illuminate starting at a beginning of the occupancy time interval and ending at the earlier of (i) a stop time respective to an end of the occupancy time interval, or (ii) a time determined when a hand sanitation event occurs; or b) controlling the noncompliance light to illuminate starting at an end of the occupancy time interval upon a determination that no hand sanitation event occurred during the occupancy time interval.
8. The hand hygiene compliance enforcement device of aspect 7, wherein the at least one annunciator further comprises a compliance light, and wherein the electronic processor is configured to execute the sanitation detection instructions to cause the electronic processor to perform at least one of: a) controlling the compliance light to illuminate starting at a time in the occupancy time interval upon a determination that a hand sanitation event occurs; b) controlling the compliance light to illuminate starting at an end of the occupancy time interval upon a determination that a hand sanitation event occurred during the occupancy time interval.
9. The hand hygiene compliance enforcement device of aspect 7, wherein the noncompliance light is illuminated by a flashing illumination.
10. The hand hygiene compliance enforcement device as in any one of aspects 1-9, wherein the at least one annunciator comprises at least one of: a light; an audio speaker; a display; or, a touchscreen display.
11. The hand hygiene compliance enforcement device as in any one of aspects 1-10, wherein the instructions are configured to lack or filter identification of a person in the sanitation zone.
12. The hand hygiene compliance enforcement device of aspect 11, wherein the instructions comprise reporting instructions readable and executable by the electronic processor, the reporting instructions that when executed by the electronic processor, cause the electronic processor to: accumulate data statistics for a plurality of successive occupancy time intervals on corresponding determinations of whether a hand sanitation event occurs and generate a compliance report on the data statistics.
13. The hand hygiene compliance enforcement device as in any one of aspects 1-12, wherein the instructions comprise reporting instructions readable and executable by the electronic processor, the reporting instructions that when executed by the electronic processor, cause the electronic processor to: identify a person in the sanitation zone during the occupancy time interval using (i) a radio frequency identification (RFID) reader that detects RFID badges, or (ii) facial recognition applied to the real time video; accumulate data statistics for a plurality of successive occupancy time intervals on corresponding determinations of whether a hand sanitation event occurs; and generate a compliance report based on the data statistics including identification of one or more persons identified in the sanitation zone during occupancy time intervals for which the corresponding determination of whether a hand sanitation event occurs is that no hand sanitation event occurred.
14. The hand hygiene compliance enforcement device as in any one of aspects 1-13, wherein the at least one video camera consists of a single video camera and the wash station zone is contained in the sanitation zone imaged by the real time video.
15. The hand hygiene compliance enforcement device as in any one of aspects 1-14 further comprising: a hand sanitation station including at least a sanitation fluid dispenser, wherein the at least one video camera is positioned such that at least a portion of the hand sanitation station coincides with the wash station zone in the real time video acquired by the at least one video camera.
16. A non-transitory storage medium storing instructions for performing a hand hygiene compliance enforcement, instructions being readable and executable by an electronic processor, wherein the instructions, when executed by the electronic processor, cause the electronic processor to: receive real time video acquired by at least one video camera in real time; analyze the real time video to determine an occupancy time interval over which a sanitation zone imaged by the real time video is occupied; determine whether a hand sanitation event occurs based on images of a wash station zone as extracted from the real time video acquired during the occupancy time interval; and control at least one annunciator to annunciate a hand hygiene compliance status based on the determination of whether a hand sanitation event occurs.
17. The non-transitory storage medium of aspect 16, wherein determination of whether the hand sanitation event occurs comprises: applying a convolutional neural network (CNN) to the images of the wash station zone as extracted from successive frames of the real time video acquired during the occupancy time interval to generate per-frame confidence values; and determining whether the hand sanitation event occurs based on the per-frame confidence values, wherein the CNN is pre-trained to generate confidence values indicating confidence that input images depict a hand sanitation event using training images of the wash station zone or a standard wash station zone that depict a hand sanitation event and training images of the wash station zone or the standard wash station zone that do not depict a hand sanitation event.
18. The non-transitory storage medium of aspect 17 wherein application of the CNN comprises: converting the image of the wash station zone as extracted from a frame of the successive frames to a tensor of channel arrays wherein each channel array represents a color channel of the image of the wash station zone; and applying the CNN to the tensor of channel arrays.
19. The non-transitory storage medium as in any one of aspects 1-17, wherein the instructions, when executed by the electronic processor, further cause the electronic processor to: accumulate data statistics for a plurality of successive occupancy time intervals on corresponding determinations of whether a hand sanitation event occurs; and generate a compliance report based on the data statistics.
20. A hand hygiene compliance enforcement method comprising: acquiring, via a at least one video camera, real time video of a field of view that includes a sanitation station having a wash station zone; determining, via an electronic processor, whether a hand sanitation event occurs based on images of the wash station zone as extracted from the real time video; and annunciating, via an annunciator, a compliance status based on the determination of whether a hand sanitation event occurs.
21. The hand hygiene compliance enforcement method of aspect 20, wherein determination of whether a hand sanitation event occurs comprises: applying a machine learning (ML) classifier to the images of the wash station zone as extracted from successive frames of the real time video to generate per-frame confidence values; and determining whether a hand sanitation event occurs based on the per-frame confidence values for the successive frames, wherein the ML classifier is pre-trained using training images of the wash station zone or a standard wash station zone that depict a hand sanitation event and training images of the wash station zone or the standard wash station zone that do not depict a hand sanitation event.
22. The hand hygiene compliance enforcement method of aspect 21, wherein application of the ML classifier comprises: converting an image of the wash station zone as extracted from the frame of the successive frames to a tensor of channel arrays, wherein each channel array represents a color channel of the image of the wash station zone; and applying the ML classifier to the tensor of channel arrays.
23. The hand hygiene compliance enforcement method as in any one of aspects 1-22 further comprising analyzing the real time video to determine an occupancy time interval over which a sanitation zone imaged by the real time video is occupied, wherein the determination of whether a hand sanitation event occurs is based on images of the wash station zone as extracted from only the real time video acquired during the occupancy time interval.
24. The hand hygiene compliance enforcement method as in any one of aspects 1-23 further comprising determining, with the electronic processor, whether a gowning event occurs based on images of a gowning zone as extracted from the real time video, wherein the compliance status is further based on the determination of whether a gowning event occurs.
25. The hand hygiene compliance enforcement method of aspect 24, wherein determination of whether the gowning event occurs comprises: applying a machine learning (ML) classifier to the images of the gowning zone as extracted from successive frames of the real time video to generate per-frame gowning event confidence values; and determining whether the gowning event occurs based on the per-frame gowning event confidence values for the successive frames, wherein the ML classifier for generating the per-frame gowning event confidence values is pre-trained using training images of the gowning zone or a standard gowning zone that depict a gowning event and training images of the gowning zone or the standard gowning zone that do not depict a gowning event.
26. A hand hygiene compliance enforcement device comprising: at least one non-camera-based sensor; at least one annunciator; an electronic processor; and a non-transitory storage medium storing instructions that are readable and executable by the electronic processor, wherein the instructions, when executed by the electronic processor, cause the electronic processor to: process real time data output of the non-camera-based sensor in real time, analyze, by execution of occupancy detection instructions, the data output to determine an occupancy time interval over which a sanitation zone determined by the data output is occupied, determine, by execution of sanitation detection instructions, whether a hand sanitation event occurs based on the data output corresponding to a wash station zone, the data output acquired during the occupancy time interval, and control the at least one annunciator to annunciate a hand hygiene compliance status based on the hand sanitation event.
27. The hand hygiene compliance enforcement device of aspect 26, wherein the electronic processor is configured to execute the sanitation detection instructions to cause the electronic processor to: apply, for each series of data output of the non-camera-based sensor as acquired during the occupancy time interval, a machine learning (ML) classifier to sensor data corresponding to the wash station zone, to generate a confidence value indicative of likelihood that the sensor data of the wash station zone corresponds a hand sanitation event, and determine whether a hand sanitation event occurs in the wash station zone based on the confidence values for the successive frames, wherein the ML classifier is pre-trained to generate the confidence value using training sensor data of the wash station zone or a standard wash station zone that defines a hand sanitation event and sensor data of the wash station zone or the standard wash station zone that do not correspond to a hand sanitation event.
28. The hand hygiene compliance enforcement device of aspect 27, wherein application of the ML classifier comprises: converting sensor data of the wash station zone to a tensor of channel arrays, wherein each channel array emulates a color channel of an image of the wash station zone; and applying the ML classifier to the tensor of channel arrays.
29. The hand hygiene compliance enforcement device of aspect 1 further comprising an information sharing application comprising an information sharing interface, the information sharing interface configured to display sanitation or hygiene compliance data or information.
30. The hand hygiene compliance enforcement device of aspect 1 further comprising a hand hygiene marketing platform comprising a marketing interface, the marketing interface configured to display advertisements to users in the sanitation zone or a marketing zone.
Additional Considerations
Although the disclosure herein sets forth a detailed description of numerous different embodiments, it should be understood that the legal scope of the description is defined by the words of the claims set forth at the end of this patent and equivalents. The detailed description is to be construed as exemplary only and does not describe every possible embodiment since describing every possible embodiment would be impractical. Numerous alternative embodiments may be implemented, using either current technology or technology developed after the filing date of this patent, which would still fall within the scope of the claims.
The following additional considerations apply to the foregoing discussion. Throughout this specification, plural instances may implement components, operations, or structures described as a single instance. Although individual operations of one or more methods are illustrated and described as separate operations, one or more of the individual operations may be performed concurrently, and nothing requires that the operations be performed in the order illustrated. Structures and functionality presented as separate components in example configurations may be implemented as a combined structure or component. Similarly, structures and functionality presented as a single component may be implemented as separate components. These and other variations, modifications, additions, and improvements fall within the scope of the subject matter herein.
Additionally, certain embodiments are described herein as including logic or a number of routines, subroutines, applications, or instructions. These may constitute either software (e.g., code embodied on a machine-readable medium or in a transmission signal) or hardware. In hardware, the routines, etc., are tangible units capable of performing certain operations and may be configured or arranged in a certain manner. In example embodiments, one or more computer systems (e.g., a standalone, client or server computer system) or one or more hardware modules of a computer system (e.g., a processor or a group of processors) may be configured by software (e.g., an application or application portion) as a hardware module that operates to perform certain operations as described herein.
In various embodiments, a hardware module may be implemented mechanically or electronically. For example, a hardware module may comprise dedicated circuitry or logic that is permanently configured (e.g., as a special-purpose processor, such as a field programmable gate array (FPGA) or an application-specific integrated circuit (ASIC)) to perform certain operations. A hardware module may also comprise programmable logic or circuitry (e.g., as encompassed within a general-purpose processor or other programmable processor) that is temporarily configured by software to perform certain operations. It will be appreciated that the decision to implement a hardware module mechanically, in dedicated and permanently configured circuitry, or in temporarily configured circuitry (e.g., configured by software) may be driven by cost and time considerations.
Accordingly, the term “hardware module” should be understood to encompass a tangible entity, be that an entity that is physically constructed, permanently configured (e.g., hardwired), or temporarily configured (e.g., programmed) to operate in a certain manner or to perform certain operations described herein. Considering embodiments in which hardware modules are temporarily configured (e.g., programmed), each of the hardware modules need not be configured or instantiated at any one instance in time. For example, where the hardware modules comprise a general-purpose processor configured using software, the general-purpose processor may be configured as respective different hardware modules at different times. Software may accordingly configure a processor, for example, to constitute a particular hardware module at one instance of time and to constitute a different hardware module at a different instance of time.
Hardware modules may provide information to, and receive information from, other hardware modules. Accordingly, the described hardware modules may be regarded as being communicatively coupled. Where multiple of such hardware modules exist contemporaneously, communications may be achieved through signal transmission (e.g., over appropriate circuits and buses) that connect the hardware modules. In embodiments in which multiple hardware modules are configured or instantiated at different times, communications between such hardware modules may be achieved, for example, through the storage and retrieval of information in memory structures to which the multiple hardware modules have access. For example, one hardware module may perform an operation and store the output of that operation in a memory device to which it is communicatively coupled. A further hardware module may then, at a later time, access the memory device to retrieve and process the stored output. Hardware modules may also initiate communications with input or output devices, and may operate on a resource (e.g., a collection of information).
The various operations of example methods described herein may be performed, at least partially, by one or more processors that are temporarily configured (e.g., by software) or permanently configured to perform the relevant operations. Whether temporarily or permanently configured, such processors may constitute processor-implemented modules that operate to perform one or more operations or functions. The modules referred to herein may, in some example embodiments, comprise processor-implemented modules.
Similarly, the methods or routines described herein may be at least partially processor-implemented. For example, at least some of the operations of a method may be performed by one or more processors or processor-implemented hardware modules. The performance of certain of the operations may be distributed among the one or more processors, not only residing within a single machine, but deployed across a number of machines. In some example embodiments, the processor or processors may be located in a single location, while in other embodiments the processors may be distributed across a number of locations.
The performance of certain of the operations may be distributed among the one or more processors, not only residing within a single machine, but deployed across a number of machines. In some example embodiments, the one or more processors or processor-implemented modules may be located in a single geographic location (e.g., within a home environment, an office environment, or a server farm). In other embodiments, the one or more processors or processor-implemented modules may be distributed across a number of geographic locations.
This detailed description is to be construed as exemplary only and does not describe every possible embodiment, as describing every possible embodiment would be impractical, if not impossible. A person of ordinary skill in the art may implement numerous alternate embodiments, using either current technology or technology developed after the filing date of this application.
Those of ordinary skill in the art will recognize that a wide variety of modifications, alterations, and combinations can be made with respect to the above described embodiments without departing from the scope of the invention, and that such modifications, alterations, and combinations are to be viewed as being within the ambit of the inventive concept.
The patent claims at the end of this patent application are not intended to be construed under 35 U.S.C. § 112(f) unless traditional means-plus-function language is expressly recited, such as “means for” or “step for” language being explicitly recited in the claim(s). The systems and methods described herein are directed to an improvement to computer functionality, and improve the functioning of conventional computers.
Claims
1. A hand hygiene compliance enforcement device comprising:
- at least one video camera;
- at least one annunciator;
- an electronic processor; and
- a non-transitory storage medium storing instructions that are readable and executable by the electronic processor, wherein the instructions, when executed by the electronic processor, cause the electronic processor to:
- process real time video acquired by the at least one video camera in real time,
- analyze, by execution of occupancy detection instructions, the real time video to determine an occupancy time interval over which a sanitation zone imaged by the real time video is occupied,
- determine, by execution of sanitation detection instructions, whether a hand sanitation event occurs based on images of a wash station zone as extracted from the real time video acquired during the occupancy time interval,
- convert an image of the wash station zone as extracted from a frame of one or more frames to a tensor of channel arrays wherein each channel array represents a color channel of the image of the wash station zone,
- apply a convolutional neural network (CNN) to the tensor of channel arrays, and
- control the at least one annunciator to annunciate a hand hygiene compliance status based on the determination of whether a hand sanitation event occurs.
2. The hand hygiene compliance enforcement device of claim 1, wherein the electronic processor is configured to execute the sanitation detection instructions to cause the electronic processor to:
- apply, for each respective frame of successive frames of the real time video acquired during the occupancy time interval, a machine learning (ML) classifier to one or more images of the wash station zone as extracted from the respective frame to generate a confidence value indicative of likelihood that the one or more images of the wash station zone depicts a hand sanitation event, and
- determine whether a hand sanitation event occurs in the wash station zone based on the confidence values for the successive frames,
- wherein the ML classifier is pre-trained to generate the confidence value using (1) training images of the wash station zone or a standard wash station zone that depicts a hand sanitation event, and (2) training images of the wash station zone or the standard wash station zone that do not depict a hand sanitation event.
3. The hand hygiene compliance enforcement device of claim 2, wherein application of the ML classifier comprises:
- converting an image of the wash station zone as extracted from a frame of the successive frames to a tensor of channel arrays, wherein each channel array represents a color channel of the image of the wash station zone; and
- applying the ML classifier to the tensor of channel arrays.
4. The hand hygiene compliance enforcement device of claim 1,
- wherein the images of the wash station zone, as extracted from the real time video acquired during the occupancy time interval, have pixel dimensions length (L) by width (W) where L is 100 pixels or less and W is 100 pixels or less, and
- wherein extraction of the images of the wash station zone from the real time video acquired during the occupancy time interval comprises pixel resampling, the pixel dimensions L by W being determined after the pixel resampling.
5. The hand hygiene compliance enforcement device of claim 1, wherein the electronic processor is configured to execute the sanitation detection instructions to cause the electronic processor to:
- determine whether a hand sanitation event occurs in the wash station zone by detecting a color of a hand sanitation fluid in the images of the wash station zone as extracted from the real time video acquired during the occupancy time interval.
6. The hand hygiene compliance enforcement device of claim 1, wherein the electronic processor is configured to execute the occupancy detection instructions to cause the electronic processor to perform at least one of:
- a) extracting, for one or more successive frames of the real time video, an extracted image of the sanitation zone from a frame of the one or more successive frames, and comparing the extracted image of the sanitation zone with a reference image of the sanitation zone to detect whether the extracted image of the sanitation zone indicates occupation of the sanitation zone; or
- b) extracting, for two or more successive frames of the real time video, an extracted image of the sanitation zone from a respective frame of the two or more successive frames, and comparing a first image of the sanitation zone as extracted from a first respective frame of the two or more successive frames with a second image of the sanitation zone as extracted from a second respective frame of the two or more successive frames to detect frame-to-frame changes that indicate occupation of the sanitation zone.
7. The hand hygiene compliance enforcement device of claim 1,
- wherein the at least one annunciator comprises a noncompliance light, and
- wherein the electronic processor is configured to execute the sanitation detection instructions to cause the electronic processor to perform at least one of:
- a) controlling the noncompliance light to illuminate starting at a beginning of the occupancy time interval and ending at the earlier of (i) a stop time respective to an end of the occupancy time interval, or (ii) a time determined when a hand sanitation event occurs; or
- b) controlling the noncompliance light to illuminate starting at an end of the occupancy time interval upon a determination that no hand sanitation event occurred during the occupancy time interval.
8. The hand hygiene compliance enforcement device of claim 7,
- wherein the at least one annunciator further comprises a compliance light, and
- wherein the electronic processor is configured to execute the sanitation detection instructions to cause the electronic processor to perform at least one of:
- a) controlling the compliance light to illuminate starting at a time in the occupancy time interval upon a determination that a hand sanitation event occurs;
- b) controlling the compliance light to illuminate starting at an end of the occupancy time interval upon a determination that a hand sanitation event occurred during the occupancy time interval.
9. The hand hygiene compliance enforcement device of claim 7, wherein the noncompliance light is illuminated by a flashing illumination.
10. The hand hygiene compliance enforcement device of claim 1, wherein the at least one annunciator comprises at least one of:
- a light;
- an audio speaker;
- a display; or
- a touchscreen display.
11. The hand hygiene compliance enforcement device of claim 1, wherein the instructions are configured to lack or filter identification of a person in the sanitation zone.
12. The hand hygiene compliance enforcement device of claim 11, wherein the instructions comprise reporting instructions readable and executable by the electronic processor, the reporting instructions that when executed by the electronic processor, cause the electronic processor to:
- accumulate data statistics for a plurality of successive occupancy time intervals on corresponding determinations of whether a hand sanitation event occurs and generate a compliance report on the data statistics.
13. The hand hygiene compliance enforcement device of claim 1, wherein the instructions comprise reporting instructions readable and executable by the electronic processor, the reporting instructions that when executed by the electronic processor, cause the electronic processor to:
- identify a person in the sanitation zone during the occupancy time interval using (i) a radio frequency identification (RFID) reader that detects RFID badges, or (ii) facial recognition applied to the real time video;
- accumulate data statistics for a plurality of successive occupancy time intervals on corresponding determinations of whether a hand sanitation event occurs; and
- generate a compliance report based on the data statistics including identification of one or more persons identified in the sanitation zone during occupancy time intervals for which the corresponding determination of whether a hand sanitation event occurs is that no hand sanitation event occurred.
14. The hand hygiene compliance enforcement device of claim 1, wherein the at least one video camera consists of a single video camera and the wash station zone is contained in the sanitation zone imaged by the real time video.
15. The hand hygiene compliance enforcement device of claim 1 further comprising:
- a hand sanitation station including at least a sanitation fluid dispenser,
- wherein the at least one video camera is positioned such that at least a portion of the hand sanitation station coincides with the wash station zone in the real time video acquired by the at least one video camera.
16. A non-transitory storage medium storing instructions for performing a hand hygiene compliance enforcement, instructions being readable and executable by an electronic processor, wherein the instructions, when executed by the electronic processor, cause the electronic processor to:
- receive real time video acquired by at least one video camera in real time;
- analyze the real time video to determine an occupancy time interval over which a sanitation zone imaged by the real time video is occupied;
- determine whether a hand sanitation event occurs based on images of a wash station zone as extracted from the real time video acquired during the occupancy time interval;
- convert an image of the wash station zone as extracted from a frame of one or more frames to a tensor of channel arrays wherein each channel array represents a color channel of the image of the wash station zone;
- apply a convolutional neural network (CNN) to the tensor of channel arrays; and
- control at least one annunciator to annunciate a hand hygiene compliance status based on the determination of whether a hand sanitation event occurs.
17. The non-transitory storage medium of claim 16, wherein determination of whether the hand sanitation event occurs comprises:
- applying the CNN to the images of the wash station zone as extracted from successive frames of the real time video acquired during the occupancy time interval to generate per-frame confidence values; and
- determining whether the hand sanitation event occurs based on the per-frame confidence values,
- wherein the CNN is pre-trained to generate confidence values indicating confidence that input images depict a hand sanitation event using training images of the wash station zone or a standard wash station zone that depict a hand sanitation event and training images of the wash station zone or the standard wash station zone that do not depict a hand sanitation event.
18. The non-transitory storage medium of claim 17, wherein the instructions, when executed by the electronic processor, further cause the electronic processor to:
- accumulate data statistics for a plurality of successive occupancy time intervals on corresponding determinations of whether a hand sanitation event occurs; and
- generate a compliance report based on the data statistics.
19. A hand hygiene compliance enforcement method comprising:
- acquiring, via a at least one video camera, real time video of a field of view that includes a sanitation station having a wash station zone;
- analyzing the real time video to determine an occupancy time interval over which the wash station zone imaged by the real time video is occupied;
- determining, via an electronic processor, whether a hand sanitation event occurs based on images of the wash station zone as extracted from the real time video;
- converting an image of the wash station zone as extracted from a frame of one or more frames to a tensor of channel arrays wherein each channel array represents a color channel of the image of the wash station zone;
- applying a convolutional neural network (CNN) to the tensor of channel arrays; and
- annunciating, via an annunciator, a compliance status based on the determination of whether a hand sanitation event occurs.
20. The hand hygiene compliance enforcement method of claim 19, wherein determination of whether a hand sanitation event occurs comprises:
- applying a machine learning (ML) classifier to the images of the wash station zone as extracted from successive frames of the real time video to generate per-frame confidence values; and
- determining whether a hand sanitation event occurs based on the per-frame confidence values for the successive frames,
- wherein the ML classifier is pre-trained using training images of the wash station zone or a standard wash station zone that depict a hand sanitation event and training images of the wash station zone or the standard wash station zone that do not depict a hand sanitation event.
21. The hand hygiene compliance enforcement method of claim 20, wherein application of the ML classifier comprises:
- converting an image of the wash station zone as extracted from the frame of the successive frames to a tensor of channel arrays, wherein each channel array represents a color channel of the image of the wash station zone; and
- applying the ML classifier to the tensor of channel arrays.
22. The hand hygiene compliance enforcement method of claim 19,
- wherein the determination of whether a hand sanitation event occurs is based on images of the wash station zone as extracted from only the real time video acquired during the occupancy time interval.
23. The hand hygiene compliance enforcement method of claim 19 further comprising determining, with the electronic processor, whether a gowning event occurs based on images of a gowning zone as extracted from the real time video,
- wherein the compliance status is further based on the determination of whether a gowning event occurs.
24. The hand hygiene compliance enforcement method of claim 23, wherein determination of whether the gowning event occurs comprises:
- applying a machine learning (ML) classifier to the images of the gowning zone as extracted from successive frames of the real time video to generate per-frame gowning event confidence values; and
- determining whether the gowning event occurs based on the per-frame gowning event confidence values for the successive frames,
- wherein the ML classifier for generating the per-frame gowning event confidence values is pre-trained using training images of the gowning zone or a standard gowning zone that depict a gowning event and training images of the gowning zone or the standard gowning zone that do not depict a gowning event.
25. A hand hygiene compliance enforcement device comprising:
- at least one non-camera-based sensor;
- at least one annunciator;
- an electronic processor; and
- a non-transitory storage medium storing instructions that are readable and executable by the electronic processor, wherein the instructions, when executed by the electronic processor, cause the electronic processor to:
- process real time data output of the non-camera-based sensor in real time, analyze, by execution of occupancy detection instructions, the data output to determine an occupancy time interval over which a sanitation zone determined by the data output is occupied,
- determine, by execution of sanitation detection instructions, whether a hand sanitation event occurs based on the data output corresponding to a wash station zone, the data output acquired during the occupancy time interval,
- convert an image of the wash station zone as extracted from a frame of one or more frames to a tensor of channel arrays wherein each channel array represents a color channel of the image of the wash station zone,
- apply a convolutional neural network (CNN) to the tensor of channel arrays, and
- control the at least one annunciator to annunciate a hand hygiene compliance status based on the hand sanitation event.
26. The hand hygiene compliance enforcement device of claim 25, wherein the electronic processor is configured to execute the sanitation detection instructions to cause the electronic processor to:
- apply, for each series of data output of the non-camera-based sensor as acquired during the occupancy time interval, a machine learning (ML) classifier to sensor data corresponding to the wash station zone, to generate a confidence value indicative of likelihood that the sensor data of the wash station zone corresponds a hand sanitation event, and
- determine whether a hand sanitation event occurs in the wash station zone based on the confidence values for the successive frames,
- wherein the ML classifier is pre-trained to generate the confidence value using training sensor data of the wash station zone or a standard wash station zone that defines a hand sanitation event and sensor data of the wash station zone or the standard wash station zone that do not correspond to a hand sanitation event.
27. The hand hygiene compliance enforcement device of claim 26, wherein application of the ML classifier comprises:
- converting sensor data of the wash station zone to a tensor of channel arrays, wherein each channel array emulates a color channel of an image of the wash station zone; and
- applying the ML classifier to the tensor of channel arrays.
28. The hand hygiene compliance enforcement device of claim 1 further comprising an information sharing application comprising an information sharing interface, the information sharing interface configured to display sanitation or hygiene compliance data or information.
29. The hand hygiene compliance enforcement device of claim 1 further comprising a hand hygiene marketing platform comprising a marketing interface, the marketing interface configured to display advertisements to users in the sanitation zone or a marketing zone.
6038331 | March 14, 2000 | Johnson |
10002518 | June 19, 2018 | Wheeler |
11290693 | March 29, 2022 | Ross |
20080019490 | January 24, 2008 | Lynn |
20080107341 | May 8, 2008 | Lu |
20090051545 | February 26, 2009 | Koblasz |
20090087028 | April 2, 2009 | Lacey |
20100117836 | May 13, 2010 | Seyed Momen |
20120212582 | August 23, 2012 | Deutsch |
20130278635 | October 24, 2013 | Maggiore |
20150109442 | April 23, 2015 | Derenne |
20160042635 | February 11, 2016 | Rosebraugh |
20160048726 | February 18, 2016 | Tang |
20160314683 | October 27, 2016 | Felch |
20180047277 | February 15, 2018 | Thyroff |
20180232602 | August 16, 2018 | Dey |
20180301014 | October 18, 2018 | Worral |
20180357886 | December 13, 2018 | Tavori |
20190095764 | March 28, 2019 | Li |
20200167715 | May 28, 2020 | Bhatt |
20200323397 | October 15, 2020 | Simonovsky |
Type: Grant
Filed: Apr 21, 2021
Date of Patent: Jun 7, 2022
Patent Publication Number: 20210241603
Assignee: VISUALQ (Downers Grove, IL)
Inventors: John Caleb Harman (Downers Grove, IL), John M. Abdelsayed (Pepper Pike, OH)
Primary Examiner: Joseph H Feild
Assistant Examiner: Rufus C Point
Application Number: 17/236,202
International Classification: G08B 21/24 (20060101); G06K 9/62 (20220101); H04N 7/18 (20060101); G08B 5/36 (20060101); G16H 40/20 (20180101); G06N 20/00 (20190101); G06K 7/10 (20060101); G06Q 30/00 (20120101); G06Q 10/10 (20120101); G16H 15/00 (20180101); G06Q 30/02 (20120101); G06V 20/40 (20220101);