Login| Sign Up| Help| Contact|

Patent Searching and Data


Title:
SYSTEMS AND METHODS FOR LIVE PROJECTION IMAGING FOR FLUORESCENCE MICROSCOPY
Document Type and Number:
WIPO Patent Application WO/2022/026952
Kind Code:
A1
Abstract:
Implementations discussed and claimed herein provide systems and methods live projection imaging for fluorescence microscopy. In one implementation, a 3D view of a sample, such as cells, is generated for direct viewing. A projection of a volume is generated that is optically sheared into a single camera frame in light-sheet fluorescence microscopy. Optical shearing is synchronized with acquisition of a volume, where volumetric information may be directly viewed in a single acquisition to evaluate cellular 3D morphologies and dynamics.

Inventors:
FIOLKA RETO (US)
Application Number:
PCT/US2021/044193
Publication Date:
February 03, 2022
Filing Date:
August 02, 2021
Export Citation:
Click for automatic bibliography generation   Help
Assignee:
UNIV TEXAS (US)
International Classes:
G02B21/36; G01N21/17; G02B21/00; G02B21/02; G02B21/06
Domestic Patent References:
WO2020058556A12020-03-26
Foreign References:
US3548730A1970-12-22
US20110115895A12011-05-19
US20130335818A12013-12-19
Attorney, Agent or Firm:
ELLIOTT, Taryn et al. (US)
Download PDF:
Claims:
CLAIMS

What is claimed is:

1. A method for imaging a sample using a microscopy system, the method comprising: scanning the sample along a scanning direction while illuminating the sample with light generated by at least one light source along an illuminating direction; shearing fluorescence light emitted by the sample in response to illumination of the sample, the fluorescence light sheared optically using a shearing unit to form a sheared volume; and projecting the sheared volume onto a single frame of an imager in real time using the shearing unit.

2. The method of claim 1, wherein the shearing unit includes a first mirror and a second mirror of a modular galvanometer.

3. The method of claim 2, further comprising: displacing the fluorescence light in a displacement direction across the imager by rotating the first mirror and the second mirror in synchrony when the sample is scanned.

4. The method of claim 3, wherein the displacement direction is lateral.

5. The method of any of claims 2-4, wherein the modular galvanometer is mounted to the imager.

6. The method of any of claims 2-5, wherein the first mirror and the second mirror are made from gold.

7. The method of any of claims 2-6, wherein the imager is a sCMOS camera or a CCD camera.

8. The method of any of claims 1-7, wherein the scanning direction is diagonal.

9. The method of any of claims 1-8, further comprising: providing a direct view of the sheared volume of the sample in real-time.

10. The method of claim 9, further comprising: adjusting at least one imaging parameter based on the direct view.

11. The method of claim 10, wherein the at least one imaging parameter includes one or more of light-sheet position, focus, beam property, alignment, and scan range.

12. The method of any of claims 9-11, wherein the sample includes cells, the direct view including cellular 3D morphologies and dynamics of the cells.

13. The method of any of claims 9-12, wherein the direct view includes a visual impression that the sample is being viewed from different angles, the visual impression generated by varying a magnitude of shearing for the sheared volume being projected.

14. The method of any of claims 1-13, wherein the microscopy system is a lattice light- sheet microscopy system.

15. The method of any of claims 1-14, wherein the microscopy system is an oblique plane microscopy system.

16. The method of any of claims 1-15, wherein the microscopy system is an epifluorescence microscopy system.

17. The method of any of claims 1-16, wherein a summation is formed by scanning the sample one or more times during exposure of the single frame.

18. The method of any of claims 1-17, wherein projecting the stacked volume includes generating a sum projection by numerically summing image tiles associated with the stacked volume.

19. The method of any of claims 1-18, wherein a speed at which the sheared volume is projected is tied to a frame rate of the imager.

20. The method of any of claims 1-19, wherein the fluorescence light is directed through a tube lens into the shearing unit.

21. The method of any of claims 1-20, wherein the fluorescence light is sheared synchronously with acquisition by the imager.

22. A system adapted to execute the method of any of claims 1-21.

23. One or more tangible non-transitory computer-readable storage media storing computer-executable instructions for performing a computer process on a computing system, the computer process comprising the method of any of claims 1-21.

Description:
PATENT COOPERATION TREATY APPLICATION

SYSTEMS AND METHODS FOR LIVE PROJECTION IMAGING FOR FLUORESCENCE MICROSCOPY

SYSTEMS AND METHODS FOR LIVE PROJECTION IMAGING FOR

FLUORESCENCE MICROSCOPY

CROSS-REFERENCE TO RELATED APPLICATIONS

[0001] The present application claims priority to U.S. Provisional Patent Applications Nos. 63/063,779, filed on August 10, 2020 and 63/059,554, filed on July 31, 2020, both entitled “SYSTEMS AND METHODS FOR LIVE PROJECTION IMAGING FOR FLUORESCENCE MICROSCOPY.” Each of these applications is specifically incorporated by reference in its entirety herein.

ACKNOWLEDGEMENT OF GOVERNMENT SUPPORT

[0002] This invention was made with government support under Grant Number GM133522 awarded by the National Institutes of Health, and Grant Number RRl 60057 awarded by the Cancer Prevention Institute of Texas. The government has certain rights in the invention.

BACKGROUND

1. Field

[0003] The present disclosure is directed to systems and methods for directly viewing a volume and more particularly to generating a projection of a three-dimensional (3D) volume in real time for fluorescence microscopy.

2. Discussion of Related Art

[0004] To understand a sample of a specimen, ranging from inanimate viruses to complex multicellular organisms, various optical techniques are used to provide non-invasive, low phototoxicity, and high resolution volumetric imaging. For example, 3D fluorescence microscopy typically involves a serial acquisition of two-dimensional (2D) images from which a 3D volume is rendered to view individual cells or aspects of the specimen. Stated differently, to make sense of 3D data for a specimen, a 3D stack including hundreds of 2D images is acquired, and a 3D image of the specimen is rendered using a graphics program. Not only is this process time intensive and computationally burdensome, but capturing images of and understanding fast biological or other dynamic processes associated with the specimen is challenging using this process. More particularly, because this process is typically not performed in real-time, it remains challenging to identify cells of interest, as well as to optimize image acquisition parameters (e.g., illumination properties, alignment, scan range, etc.). Furthermore, it limits the volumetric image acquisition rate, adds substantial computational overhead, results in duplicated data, and ultimately diminishes the throughput of a microscope generally used in connection with this process. It is with these observations in mind, among others, that various aspects of the present disclosure were conceived and developed.

BRIEF DESCRIPTION OF THE DRAWINGS

[0005] To describe the manner in which the advantages and features of the disclosure can be obtained, reference is made to implementations thereof which are illustrated in the appended drawings. One of skill in the art will understand that the reference numbers in the following figures are repeated throughout Figures 1-10 so as to refer to the same or substantially the same features. Understanding that these drawings depict only exemplary implementations of the disclosure and are not therefore to be considered to be limiting of its scope, the principles herein are described and explained with additional specificity and detail through the use of the accompanying drawings.

[0006] Figure 1 is a diagram showing an example system for live projection imaging using lattice light-sheet microscopy or oblique plane microscopy.

[0007] Figures 2A-2B are a perspective view and a top view, respectively, of an example galvanometric shearing unit attached to a camera.

[0008] Figure 3 depicts an example system for live projection imaging using epifluorescence microscopy.

[0009] Figures 4A and 4B illustrate projection of an image, without and with using a lateral shearing unit, respectively.

[0010] Figure 5 shows a series of example images comparing a stacked volume and a live projection on a single camera frame.

[0011] Figure 6 illustrates blebs on an example image of a cell and associated kymographs.

[0012] Figure 7 illustrates an example live projection of cultured cortical neurons, including magnified views at different time points.

[0013] Figure 8 shows a signal trace from a region shown in one of the magnified views of Figure 7.

[0014] Figure 9 illustrates a comparison between a projection from a sheared image and a projection from a rotated image.

[0015] Figure 10 is an example computing system that may implement various systems and methods discussed herein. DETAILED DESCRIPTION

[0016] It will be appreciated that numerous specific details are set forth in order to provide a thorough understanding of the implementations described herein. However, it will be understood by those of ordinary skill in the art that the embodiments described herein can be practiced without these specific details. In other instances, methods, procedures and components have not been described in detail so as not to obscure the related relevant feature being described. Also, the description is not to be considered as limiting the scope of the implementations described herein.

[0017] Aspects of the present disclosure generally involve systems and methods for generating a 3D view of a sample, such as cells, for direct viewing. In one aspect, a direct, 3D view of a volume is provided using sensitive microscopy, such as lattice light-sheet microscopy (LLSM), oblique plane microscopy (OPM), or other light-sheet fluorescence microscopy (LSFM). As opposed to acquiring a 3D stack comprising sometimes hundreds of 2D images for rendering in a graphics program, the presently disclosed technology renders a 3D volume optically and in real-time (i.e., instantaneously or near instantaneously). More particularly, in one aspect, the presently disclosed technology generates a projection of a volume that is optically sheared into a single camera frame in light-sheet fluorescence microscopy, such as LLSM, OPM, and/or the like. Optical shearing is synchronized with acquisition of a volume, where volumetric information may be directly viewed in a single acquisition to evaluate cellular 3D morphologies and dynamics. Overall, the presently disclosed technology provides ultrafast imaging of volumes, thereby facilitating capture of fast biological processes, reducing computational burden of image acquisition, increasing microscopy throughput, and simplifying navigation and evaluation of a 3D sample, among other advantages.

[0018] Overall, the various systems and methods of the presently disclosed technology provide a shearing unit deployed in an optical system that displaces images on a camera to generate an effect of a projection under different viewing angles. Example implementations discussed herein involve LLSM, OPM, and similar light-sheet microscopy. However, it will be appreciated by those skilled in the art that the presently disclosed technology is applicable to any microscope forming an image on a camera and permitting rapid change of the focal plane being viewed. As one example, the systems and methods described herein may be applicable to spinning disk microscopy. However, other classes of microscopes are contemplated as well. Independent of the class of microscopy, by using the shearing unit of the presently disclosed technology, a real-time projection of a 3D sample is formed, and the viewing angle of the projection is adjustable by changing an amount of shearing.

[0019] To begin a detailed description of an example system 100 for live projection imaging, reference is made to Figure 1. In one implementation, the system 100 utilizes LSFM techniques. The highest performance subcellular LSFM techniques, including OPM and LLSM, image adherent cells that are oriented diagonally to the illumination and detection axes of the system 100, such that a sample 102 is scanned in a diagonal direction relative to a direction of illumination propagation and a direction of fluorescence detection. Generally, LSFM techniques utilize short, thin light-sheets, thereby providing improved z-resolution and reduced coverslip induced aberrations. However, as described herein, LSFM techniques often involve the acquisition of hundreds of image planes per cell, all of which being computationally sheared to permit visualization and quantification. Conversely, in one implementation, the system 100 performs shearing optically as opposed to computationally. The system 100 projects a sheared volume onto a single camera frame, permitting a 3D rendering of the sample 102 to be viewed in real-time prior to volumetric imaging. Alternatively or additionally, the system 100 permits rapid cellular process of the sample 102 to be volumetrically imaged in a 2D projection format with uncompromised lateral resolution.

[0020] In one implementation, the sample 102 is disposed on a coverslip 104 relative to an objective 108. The system 100 scans the sample 102, and a path of fluorescence light 108 is directed through a lens 110 (e.g., a tube lens) into an imager, such as a camera 120. The camera 120 may be a sCMOS, CCD, and/or similar camera or imager. Further, it will be appreciated that various types of imagers may be used. In alternative or addition to the camera 120, an eyepiece may be used, as described herein, for viewing the sample 102 with the naked eye.

[0021] As can be understood from Figures 1-2B, in one implementation, a shearing unit 114 is disposed in front of the camera 120. For example, the shearing unit 114 may attach to a c- mount of a scientific camera. The shearing unit 114 may include a first mirror 116 and a second mirror 118. The mirrors 116-118 may be galvanometric mirrors coated with gold or similar material, such that the shearing unit 114 is a modular galvanometer-based scan unit that attaches to the camera 120 to perform shearing synchronously with acquisition of a z-stack. As the sample 102 is scanned, the fluorescence light 108 is swept across the camera 120, permitting volumetric information of the sample 102 to be analyzed and viewed directly in a single acquisition. Thus, the speed at which the volumetric information is obtained is tied to the frame rate of the camera 120, increasing acquisition speed over conventional systems by 200 times or more. While the shearing unit 114 is described as including a set of galvanometric mirrors in this example, it will be appreciated by those skilled in the art that translation of the images may be performed by a variety of mechanisms. For example, the shearing unit 114 may include one galvanometric mirror and two scan lenses, and/or the like.

[0022] The system 100 scans the sample 102 along a scanning direction 126. In the example shown in Figure 1, the scanning direction 126 is diagonal compared a detection direction 122 and an illumination direction 124. The system 100 scans the sample 102 oriented on the coverslip 104 along the scanning direction 126 while illuminating the sample 102 with a light- sheet generated by at least one light source along the illumination direction 124. In one implementation, when the sample 102 is scanned, the mirrors 116 and 118 are rotated in synchrony. The synchronized rotation of the mirrors 116-118 of the shearing unit 114 displaces the image created from the fluorescence light 108 across the camera 120 in a displacement direction 106. In one implementation, the displacement direction 106 is lateral, such that the image is displaced laterally across the camera 120.

[0023] Referring to Figure 4A, in the absence of the shearing unit 114, the sample 102 is scanned along the scanning direction 126 on a diagonal trajectory relative to the detection direction 122, while being illuminated with a light-sheet in the illumination direction 124. As shown in Figure 4A, the images 202-204 (images 1... n) acquired are sequentially stacked in a stacking direction 210 in the z-dimension to form a 3D stack. The 3D stack does not reflect the x-component of the scan trajectory along the scanning direction 126, only the axial- component along the illumination direction 124. As such, if the 3D stack is viewed along a viewing direction 206 as a projection 200, the 3D stack will appear distorted.

[0024] Referring to Figure 4B, when the sample 102 is scanned, the shearing unit 114 causes the image to be displaced laterally on the camera 120, as described herein. The images 202- 204 acquired are stacked in the stacking direction 210 in the z-dimension and laterally shifted in a shearing direction 208 along the x-dimension. Thus, the images 202-204 are stacked according to the natural positions of the sections of the sample 102 along the scanning direction 126 in sample space and in image space. With the images 202-204 being located in the corresponding natural positions, the projection 200 viewed along the viewing direction 206 will appear distortion free.

[0025] Independent of whether the projection 200 appears distorted (e.g., in the example of Figure 4A) or distortion free (e.g., in the example of Figure 4B), the projection 200 may be computed by numerically summing all the tiles of the images 202-204 in the 3D stack together to generate a sum projection. In another implementation, a summation may be formed by scanning the sample 102 one or more times during a single exposure of the camera 120, such that the sum of all the images 202-204 directly appears on the camera 120 as the projection 200.

[0026] In some implementations, the projection 200 is formed such that it can be viewed with the naked eye using an eyepiece. More particularly, an eyepiece may be included with the system 100 in place of or in addition to the camera 120. The projection 200 is generated as detailed herein, and the eyepiece is disposed relative to an image plane of the projection 200 for direct view. In these implementations, a speed at which the sample 102 is scanned and a shearing speed of the shearing unit 114 are optimized to correspond to physiological limitations of the human eye. For example, the system 100 may scan the sample 102 faster than or equivalent to a video rate.

[0027] Referring to Figures 3 and 4A-4B, in one implementation, the system 100 involves epifluorescence microscopy. The sample 102 is disposed on the coverslip 104 with a focal plane 128 relative to the objective 108. In one implementation, the sample 102 or the objective 108 is scanned rapidly in a scanning direction 130 to form a 3D stack. Again, referring to Figure 4A, in the absence of the shearing unit 114, the images 202-204 acquired are sequentially stacked in the stacking direction 210 in the z-dimension to form a 3D stack. As such, if the 3D stack is viewed in the viewing direction 206, the projection 200 will appear along the optical axis. In another implementation, referring again to Figure 4B, when the sample 102 is scanned, the shearing unit 114 causes the image to be displaced laterally on the camera 120, as described herein. The images 202-204 acquired are stacked in the stacking direction 210 in the z-dimension and laterally shifted in the shearing direction 208 along the x- dimension. Regardless of whether the shearing unit 114 is utilized, the projection 200 may be computed by numerically summing all the tiles of the images 202-204 in the 3D stack together to generate a sum projection. In another implementation, a summation may be formed by scanning the sample 102 one or more times during a single exposure of the camera 120, such that the sum of all the images 202-204 directly appears on the camera 120 as the projection 200. When the shearing unit 114 is utilized, the proj ection 200 appears from a different viewing angle, permitting direct real-time views of the sample 102 that are adjustable.

[0028] Turning to Figure 5, a series 300 of example images of mammalian cells comparing a stacked volume and a live projection on a single camera frame are shown. The images 302 and 304 illustrate data volumes in LLSM or OPM before and after shearing, with the images 304 showing the projection of the data. Image 306 is a projection of a z-stack formed from 334 planes. The image 306 shows MV3 cells expressing AKtPH-GFP. As such, generation of the image 306 involved the acquisition of 334 image planes per cell, with each of the image planes being computationally sheared for visualization. Stated differently, the cells were imaged, computationally sheared, and sum intensity projected to form the image 306 showing a population of MV3 cells expressing AKtPH-GFP. Conversely, image 308 is a live projection view of the same MV3 cells expressing AKtPH-GFP acquired on a single camera frame. As can be understood from Figure 5, the image 306 and the image 308 are largely indistinguishable, despite the image 308 being acquired in a single camera frame with imaging time and computational overhead being reduced approximately hundredfold. Given that the data is sum intensity projected, information along the third axis may be sacrificed. However, unlike widefield and extended depth-of-focus microscopy, an entirety of the axial extent of the cell remains in focus, with no appreciable loss in lateral resolution and by acquiring a series of projections of varying shear amplitude, some 3D information can be encoded and potentially extracted. Further, with the light-sheet illumination, the data have high optical contrast.

[0029] As described herein, the system 100 may be used for rapid projection imaging. As an example 400 of such rapid projection imaging, turning to Figure 6, images 402-404 show blebs on another MV3 cell labeled with AKtPH-GFP, and images 406 and 408 are kymographs taken along the lines 410-412, respectively. As another example 500, referring to Figure 7, a time series of calcium waves in primary cardiomyocytes and neurons are acquired. Image 502 shows a live projection of cultured cortical neurons expressing calcium indicator GCaMP6f, with images 504 and 506 being a magnification of the region 508 of the image 502 at two different time points. Figure 8 shows a signal trace 600 from the region 510 of the image 506. As can be understood from the example 500, the ability to image volume projections directly using the present technology permits observation of calcium waves propagating through dendritic arbors that otherwise would likely have been lost if the same volume had been acquired through the serial acquisition of image planes that are computationally sheared.

[0030] Referring to Figure 9, equivalence between a projection from a rotated volume and a sheared volume is illustrated. In an example 700, an original 2D image 702 is shown. A rotated image 704 is a numerical simulation of a rotation of the original 2D image 702. A subsequent sum projection 708 of the rotated image 704 is shown. A sheared image 706, in comparison, is a numerical shearing of the original 2D image 702, and a subsequent sum projection 710 of the sheared image 706 is shown. A graph 712 comparing the sum projections 708 and 710 is shown. To compare the line profiles of the sum projections 708-710, the sum projection 710 was compressed to a same length as the sum projection 708. As shown in the graph 800, the sum projections 708-710 are largely indistinguishable with any small differences likely to have arisen from numerical errors associated with the shearing and rotation.

[0031] In one implementation, where a magnitude of shearing is varied for each projected image, a visual impression is generated that the sample 102 is being viewed from different perspectives. For LLSM and OPM applications, the shearing unit 114 performs shearing such that the lateral displacement occurring during the scanning process of the system 100 is undone. As such, the resulting projection is generally equivalent to the numerically sheared and projected 3D data coming from the system 100. However, in some implementations, the shearing unit 114 may perform a larger or smaller shearing angle than that driven by the system 100. As such, the system 100 provides shearing by an arbitrary angle and subsequent projection.

[0032] The shearing unit 114 of a live projection module may easily augment existing LLSM and OPM imaging systems. The system is compact shearing unit 114 is compact, introduces negligible light-losses (i.e., <1 %), and is fully compatible with normal microscope operation by simply switching it off. In implementations where axial information can be discarded, the system 100 allows for high-contrast and high-resolution sum projection imaging that is /V-fold faster than a traditional light-sheet microscope, where N is the number of Z-slices to acquire a 3D image stack. Despite sacrificing the spatial information in one dimension, such projection imaging of the system 100 may be useful in fields, such as neuroscience, involving rapid imaging of large volumes to resolve coordinated activity of neuronal circuits.

[0033] Generally, the system 100 acquires a 3D stack where each image is shifted laterally, such that the volume may be directly viewed as raw data. This live projection of the imaging volume may be obtained from one or more z-scans being synchronized with optical shearing during an exposure of a single frame of the camera 120. The live projection imaging may be used to explore cellular 3D morphologies and dynamics, with optimization of imaging parameters, such as light-sheet position and focus being readily adjustable during the live projection. The presently disclosed technology is able to explore such morphologies, dynamics, and fast biological processes as it is faster by a factor equaling the number of z- slices within a sequentially acquired stack. Additionally, the presently disclosed technology is simple to implement as the shearing unit 114 is a compact unit that attaches to the camera 120, such that user experience and volumetric imaging capabilities are concurrently improved.

[0034] Referring to Figure 10, a detailed description of an example computing system 900 having one or more computing units that may implement various systems and methods discussed herein is provided. The computing system 900 may be applicable to the system 100 and/or other computing or network devices for imaging, analysis, and exploration of samples. It will be appreciated that specific implementations of these devices may be of differing possible specific computing architectures not all of which are specifically discussed herein but will be understood by those of ordinary skill in the art.

[0035] The computer system 900 may be a computing system is capable of executing a computer program product to execute a computer process. Data and program files may be input to the computer system 900, which reads the files and executes the programs therein. Some of the elements of the computer system 900 are shown in Figure 10, including one or more hardware processors 902, one or more data storage devices 904, one or more memory devices 908, and/or one or more ports 908-910. Additionally, other elements that will be recognized by those skilled in the art may be included in the computing system 900 but are not explicitly depicted in Figure 10 or discussed further herein. Various elements of the computer system 900 may communicate with one another by way of one or more communication buses, point- to-point communication paths, or other communication means not explicitly depicted in Figure 10

[0036] The processor 902 may include, for example, a central processing unit (CPU), a microprocessor, a microcontroller, a digital signal processor (DSP), and/or one or more internal levels of cache. There may be one or more processors 902, such that the processor 902 comprises a single central-processing unit, or a plurality of processing units capable of executing instructions and performing operations in parallel with each other, commonly referred to as a parallel processing environment.

[0037] The computer system 900 may be a conventional computer, a distributed computer, or any other type of computer, such as one or more external computers made available via a cloud computing architecture. The presently described technology is optionally implemented in software stored on the data stored device(s) 904, stored on the memory device(s) 906, and/or communicated via one or more of the ports 908-910, thereby transforming the computer system 900 in Figure 10 to a special purpose machine for implementing the operations described herein. Examples of the computer system 900 include personal computers, terminals, workstations, mobile phones, tablets, laptops, personal computers, multimedia consoles, gaming consoles, set top boxes, and the like.

[0038] The one or more data storage devices 904 may include any non-volatile data storage device capable of storing data generated or employed within the computing system 900, such as computer executable instructions for performing a computer process, which may include instructions of both application programs and an operating system (OS) that manages the various components of the computing system 900. The data storage devices 904 may include, without limitation, magnetic disk drives, optical disk drives, solid state drives (SSDs), flash drives, and the like. The data storage devices 904 may include removable data storage media, non-removable data storage media, and/or external storage devices made available via a wired or wireless network architecture with such computer program products, including one or more database management products, web server products, application server products, and/or other additional software components. Examples of removable data storage media include Compact Disc Read-Only Memory (CD-ROM), Digital Versatile Disc Read-Only Memory (DVD- ROM), magneto-optical disks, flash drives, and the like. Examples of non-removable data storage media include internal magnetic hard disks, SSDs, and the like. The one or more memory devices 906 may include volatile memory (e.g., dynamic random access memory (DRAM), static random access memory (SRAM), etc.) and/or non-volatile memory (e.g., read only memory (ROM), flash memory, etc.).

[0039] Computer program products containing mechanisms to effectuate the systems and methods in accordance with the presently described technology may reside in the data storage devices 904 and/or the memory devices 906, which may be referred to as machine-readable media. It will be appreciated that machine-readable media may include any tangible non- transitory medium that is capable of storing or encoding instructions to perform any one or more of the operations of the present disclosure for execution by a machine or that is capable of storing or encoding data structures and/or modules utilized by or associated with such instructions. Machine-readable media may include a single medium or multiple media (e.g., a centralized or distributed database, and/or associated caches and servers) that store the one or more executable instructions or data structures.

[0040] In some implementations, the computer system 900 includes one or more ports, such as an input/output (I/O) port 908 and a communication port 910, for communicating with other computing, network, or vehicle devices. It will be appreciated that the ports 908-910 may be combined or separate and that more or fewer ports may be included in the computer system 900.

[0041] The I/O port 908 may be connected to an I/O device, or other device, by which information is input to or output from the computing system 900. Such I/O devices may include, without limitation, one or more input devices, output devices, and/or environment transducer devices.

[0042] In one implementation, the input devices convert a human-generated signal, such as, human voice, physical movement, physical touch or pressure, and/or the like, into electrical signals as input data into the computing system 900 via the I/O port 908. Similarly, the output devices may convert electrical signals received from computing system 900 via the I/O port 908 into signals that may be sensed as output by a human, such as sound, light, and/or touch. The input device may be an alphanumeric input device, including alphanumeric and other keys for communicating information and/or command selections to the processor 902 via the I/O port 908. The input device may be another type of user input device including, but not limited to: direction and selection control devices, such as a mouse, a trackball, cursor direction keys, a joystick, and/or a wheel; one or more sensors, such as a camera, a microphone, a positional sensor, an orientation sensor, a gravitational sensor, an inertial sensor, and/or an accelerometer; and/or a touch-sensitive display screen (“touchscreen”). The output devices may include, without limitation, a display, a touchscreen, a speaker, a tactile and/or haptic output device, and/or the like. In some implementations, the input device and the output device may be the same device, for example, in the case of a touchscreen.

[0043] The environment transducer devices convert one form of energy or signal into another for input into or output from the computing system 900 via the I/O port 908. For example, an electrical signal generated within the computing system 900 may be converted to another type of signal, and/or vice-versa. In one implementation, the environment transducer devices sense characteristics or aspects of an environment local to or remote from the computing device 900, such as, light, sound, temperature, pressure, magnetic field, electric field, chemical properties, physical movement, orientation, acceleration, gravity, and/or the like. Further, the environment transducer devices may generate signals to impose some effect on the environment either local to or remote from the example computing device 900, such as, physical movement of some object (e.g., a mechanical actuator), heating or cooling of a substance, adding a chemical substance, and/or the like.

[0044] In one implementation, a communication port 910 is connected to a network by way of which the computer system 900 may receive network data useful in executing the methods and systems set out herein as well as transmitting information and network configuration changes determined thereby. Stated differently, the communication port 910 connects the computer system 900 to one or more communication interface devices configured to transmit and/or receive information between the computing system 900 and other devices by way of one or more wired or wireless communication networks or connections. Examples of such networks or connections include, without limitation, Universal Serial Bus (USB), Ethernet, Wi-Fi, Bluetooth®, Near Field Communication (NFC), Long-Term Evolution (LTE), and so on. One or more such communication interface devices may be utilized via the communication port 910 to communicate one or more other machines, either directly over a point-to-point communication path, over a wide area network (WAN) (e.g., the Internet), over a local area network (LAN), over a cellular (e.g., third generation (3G) or fourth generation (4G)) network, or over another communication means. Further, the communication port 910 may communicate with an antenna or other link for electromagnetic signal transmission and/or reception.

[0045] In an example implementation, images, image data, and software and other modules and services may be embodied by instructions stored on the data storage devices 904 and/or the memory devices 906 and executed by the processor 902.

[0046] The system set forth in Figure 10 is but one possible example of a computer system that may employ or be configured in accordance with aspects of the present disclosure. It will be appreciated that other non-transitory tangible computer-readable storage media storing computer-executable instructions for implementing the presently disclosed technology on a computing system may be utilized.

[0047] In the present disclosure, the methods disclosed may be implemented as sets of instructions or software readable by a device. Further, it is understood that the specific order or hierarchy of steps in the methods disclosed are instances of example approaches. Based upon design preferences, it is understood that the specific order or hierarchy of steps in the method can be rearranged while remaining within the disclosed subject matter. The accompanying method claims present elements of the various steps in a sample order, and are not necessarily meant to be limited to the specific order or hierarchy presented.

[0048] The described disclosure may be provided as a computer program product, or software, that may include a non-transitory machine-readable medium having stored thereon instructions, which may be used to program a computer system (or other electronic devices) to perform a process according to the present disclosure. A machine-readable medium includes any mechanism for storing information in a form (e.g., software, processing application) readable by a machine (e.g., a computer). The machine-readable medium may include, but is not limited to, magnetic storage medium, optical storage medium; magneto-optical storage medium, read only memory (ROM); random access memory (RAM); erasable programmable memory (e.g., EPROM and EEPROM); flash memory; or other types of medium suitable for storing electronic instructions.

[0049] While the present disclosure has been described with reference to various implementations, it will be understood that these implementations are illustrative and that the scope of the present disclosure is not limited to them. Many variations, modifications, additions, and improvements are possible. More generally, embodiments in accordance with the present disclosure have been described in the context of particular implementations. Functionality may be separated or combined in blocks differently in various embodiments of the disclosure or described with different terminology. These and other variations, modifications, additions, and improvements may fall within the scope of the disclosure as defined in the claims that follow.