Login| Sign Up| Help| Contact|

Patent Searching and Data


Title:
METHOD AND APPARATUS FOR DENTAL VIRTUAL MODEL BASE
Document Type and Number:
WIPO Patent Application WO/2016/030754
Kind Code:
A1
Abstract:
Method and/or apparatus embodiments can provide a virtual orthodontic base, adapted to each virtual 3D dental model of the teeth (e.g., upper and lower jaw). The virtual orthodontic base should match each 3D model (e.g., size), conform to prescribed dimensional requirements and still be rapidly calculated. The 3D virtual teeth model can be provided or reconstructed (e.g., from a laser scanning, intra oral camera scanning or x-ray scan of a plaster, a negative impression (e.g., alginate or silicon) of a patient's dentition or directly from the patient's mouth).

Inventors:
REYNARD DELPHINE (FR)
HENRY SÉBASTIEN (FR)
CAPRON-RICHARD SABRINA (FR)
Application Number:
PCT/IB2015/001840
Publication Date:
March 03, 2016
Filing Date:
August 28, 2015
Export Citation:
Click for automatic bibliography generation   Help
Assignee:
TROPHY (FR)
International Classes:
A61C9/00; A61C13/00
Foreign References:
US20140186796A12014-07-03
US20020081554A12002-06-27
US20120276502A12012-11-01
Other References:
None
Attorney, Agent or Firm:
SANTARELLI (Paris, Paris, FR)
Download PDF:
Claims:
WHAT IS CLAIMED IS:

1. A computer-implemented method for generating a digital model base from a digital model of a patient's dentition, the method comprising:

forming a initial virtual base from a plurality of prescribed dimensions of a selected virtual base type;

obtaining a 3-D digital model of the patient's dentition;

determining a bounding dimensions of the 3-D digital model of the patient's dentition;

performing a first translation of the initial virtual base based on the bounding dimensions;

performing a second translation of the translated virtual base based on local criteria to form the digital model base; and

displaying, storing or transmitting the at least one virtual model base.

2. The method of claim 1 , where the selected virtual base type is a regulated virtual base type or complies with the American Board of

Orthodontics regulations.

3. The method of claim 1 , where the bounding dimensions include a 3-D bounding box of at least one of portions of an upper dental arch, portions of a lower dental arch, or both the upper dental arch and the lower dental arch in occlusion.

4. The method of claim 1 , where the plurality of prescribed dimensions of a selected virtual base type include (i) length of an oblique face of at least one pair of sides being a rear corner side, (ii) an angle of the rear corner side to a back side, (iii) an angle of the a second pair of sides being the front side to an orthogonal line from the back side; a range of height encompassing the distance form a top surface of an upper virtual base to a bottom surface of a lower virtual base, and a height limit of the upper virtual base and the lower virtual base.

5. The method of claim 1 , where the first translation of the initial virtual base comprises a global translation of the initial virtual base scaled from front to back depth dimension and a left to right width dimension.

6. The method of claim 1 , where an upper virtual base comprises a perimeter of a back side, a pair of oblique rear sides at opposite ends of the back side, a pair of front sides opposing the back side, and a pair of middle sides connecting the front sides to the rear sides.

7. The method of claim 1 , where a lower virtual base comprises a perimeter of a back side, a pair of oblique rear sides at opposite ends of the back side, a pair of middle sides opposing the back side, and a pair of middle sides respectively connected to a first end of each of the middle sides and a front curved side respectively connected between to a second end of the middle sides, where the front curved side comprises an arc of a circle.

8. The method of claim 1 , where the local criteria of the second translation of the translated virtual base comprise local movement of selected points on each o f rear sides, middle sides and front sides of the virtual base.

9. The method of claim 1 , where dental frenulems are represented on an upper virtual base and a lower virtual base.

10. The method of claim 1 , where an operator instruction can adjust a portion of the virtual model base.

1 1. The method of claim 1 , further comprising adjusting each point of the surface of the virtual model base according to z scale displacement and local xy displacement.

12. A computer-implemented method for generating a digital model base from a digital model of a patient's dentition, the method comprising:

obtaining a plurality of prescribed dimensions of a selected virtual base type;

obtaining a 3-D digital model of the patient's dentition;

determining a bounding dimensions of the 3-D digital model of the patient's dentition;

automatically forming the digital model base from the plurality of prescribed dimensions and the bounding dimensions of the 3-D digital model of the patient's dentition; and

displaying, storing or transmitting the at least one virtual model base with the 3-D digital model of the patient's dentition.

13. The method of claim 12, where the virtual model base complies with regulations of the American Board of Orthodontics.

14. The method of claim 12, where the bounding dimensions include a 3-D bounding box of at least one of portions of an upper dental arch, portions of a lower dental arch, or both the upper dental arch and the lower dental arch in occlusion.

15. The method of claim 12, where the plurality of prescribed dimensions of a selected virtual base type include at least two of (i) length of an oblique face of at least one pair of sides being a rear corner side, (ii) an angle of the rear corner side to a back side, (iii) an angle of the a second pair of sides being the front side to an orthogonal line from the back side; a range of height encompassing the distance form a top surface of an upper virtual base to a bottom surface of a lower virtual base, and a height limit of the upper virtual base and the lower virtual base.

16. An apparatus, comprising:

means for obtaining a plurality of prescribed dimensions of a selected virtual base type;

means for obtaining a 3-D digital model of the patient's dentition; means for determining a bounding dimensions of the 3-D digital model of the patient's dentition;

means for forming the digital model base from the plurality of prescribed dimensions and the bounding dimensions of the 3-D digital model of the patient's dentition; and

means for displaying, storing or transmitting the at least one virtual model base with the 3-D digital model of the patient's dentition.

Description:
METHOD AND APPARATUS FOR DENTAL VIRTUAL MODEL BASE

TECHNICAL FIELD

The disclosure relates generally to the field of dental diagnostic imaging and more particularly relates to three-dimensional imaging of patient dentition and three-dimensional bases for positioning three-dimensional teeth models.

BACKGROUND

A virtual teeth model is received, provided, or reconstructed (e.g., from a laser scanning or intra oral camera scanning of a plaster, a negative impression (e.g., alginate or silicon) of a patient's dentition or directly the patient's mouth). The virtual teeth model can also be obtained from an x-ray scan of a plaster model or a negative impression of the patient's teeth. The virtual models are positioned on a virtual base. There is, however, a need for improved methods and/or apparatus for generation of virtual teeth base models.

SUMMARY OF THE INVENTION

An aspect of this application is to advance the art of medical diagnostic treatment, particularly for dental applications.

Another aspect of this application is to address, in whole or in part, at least the foregoing and other deficiencies in the related art.

It is another aspect of the application to provide, in whole or in part, at least the advantages described herein.

An advantage offered by apparatus and/or method embodiments of the application relates to repeatable, consistent, and/or accurate positioning of a dental virtual model.

Another advantage offered by apparatus and/or method embodiments of the application relates to virtual definition of a base of a dental virtual model.

Another advantage offered by apparatus and/or method embodiments of the application relates to consistent positioning of virtual models of the teeth of a patient in accordance with dental regulatory requirement standards.

According to one aspect of the disclosure, there is provided a method for generating a digital model base from a digital model of a patient's dentition, that can include obtaining a plurality of prescribed dimensions of a selected virtual base type; obtaining a 3-D digital model of the patient's dentition; determining a bounding dimensions of the 3-D digital model of the patient's dentition; automatically forming the digital model base from the plurality of prescribed dimensions and the bounding dimensions of the 3-D digital model of the patient's dentition; and displaying, storing or transmitting the at least one virtual model base with the 3-D digital model of the patient's dentition.

These objects are given only by way of illustrative example, and such objects may be exemplary of one or more embodiments of the invention. Other desirable objectives and advantages inherently achieved by the may occur or become apparent to those skilled in the art. The invention is defined by the appended claims.

BRIEF DESCRIPTION OF THE DRAWINGS

The foregoing and other objects, features, and advantages of the invention will be apparent from the following more particular description of the embodiments of the invention, as illustrated in the accompanying drawings.

The elements of the drawings are not necessarily to scale relative to each other. Some exaggeration may be necessary in order to emphasize basic structural relationships or principles of operation. Some conventional components that would be needed for implementation of the described embodiments, such as support components used for providing power, for packaging, and for mounting and protecting x-ray system components, for example, are not shown in the drawings in order to simplify description.

Figure 1 is a schematic diagram that shows components of an imaging apparatus for surface contour imaging of a patient's teeth and related structures. Figure 2 shows schematically how patterned light is used for obtaining surface contour information using a handheld camera or other portable imaging device.

Figure 3 shows an example of surface imaging using a pattern with multiple lines of light.

Figure 4 shows a point cloud generated from structured light imaging, such as that shown in Figure 3.

Figure 5 shows a polygon mesh in the simple form of a triangular mesh.

Figure 6 is a logic flow diagram that shows an exemplary method embodiment for virtual teeth base generation according to the application.

Figures 7a-7e are diagram that show a perspective view, top view and side views of an exemplary virtual dental base model embodiment.

Figure 8 is a diagram that shows exemplary boundary dimensions to surround a 3-D teeth model.

Figure 9 is a diagram that shows exemplary sizing applied to a virtual base model embodiment.

Figure 10 is a diagram that shows exemplary global transformations applied to a virtual base model embodiment.

Figure 1 1 is a diagram that shows an exemplary local transformations applied to a virtual base model embodiment.

Figure 12 is a diagram that shows exemplary parameters for a virtual base model determination embodiment.

Figure 13 is a diagram that shows an exemplary virtual base model comprising a polygonal mesh.

Figure 14 is a diagram that illustrates local movements along an exemplary virtual base model embodiment.

DESCRIPTION OF EXEMPLARY EMBODIMENTS

The following is a description of exemplary embodiments, reference being made to the drawings in which the same reference numerals identify the same elements of structure in each of the several figures. Where they are used, the terms "first", "second", and so on, do not necessarily denote any ordinal or priority relation, but may be used for more clearly distinguishing one element or time interval from another.

The term "in signal communication" as used in the application means that tvvo or more devices and/or components are capable of communicating with each other via signals that travel over some type of signal path. Signal communication may be wired or wireless. The signals may be communication, power, data, or energy signals which may communicate information, power, and/or energy from a first device and/or component to a second device and/or component along a signal path between the first device and/or component and second device and/or component. The signal paths may include physical, electrical, magnetic, electromagnetic, optical, wired, and/or wireless connections between the first device and/or component and second device and/or component. The signal paths may also include additional devices and/or components between the first device and/or component and second device and/or component.

In the context of the present disclosure, the terms "pixel" and "voxel" may be used interchangeably to describe an individual digital image data element, that is, a single value representing a measured image signal intensity. Conventionally an individual digital image data element is referred to as a voxel for 3-dimensional or volume images and a pixel for 2-dimensional (2-D) images. For the purposes of the description herein, the terms voxel and pixel can generally be considered equivalent, describing an image elemental datum that is capable of having a range of numerical values. Voxels and pixels have attributes of both spatial location and image data code value.

"Patterned light" is used to indicate light that has a predetermined spatial pattern, such that the light has one or more features such as one or more discernible parallel lines, curves, a grid or checkerboard pattern, or other features having areas of light separated by areas without illumination. In the context of the present disclosure, the phrases "patterned light" and "structured light" are considered to be equivalent, both used to identify the light that is projected onto the head of the patient in order to derive contour image data. In the context of the present disclosure, the terms "viewer", "operator", and "user" are considered to be equivalent and refer to the viewing practitioner, technician, or other person who views and manipulates a contour image that is formed from a combination of multiple structured light images on a display monitor.

A "viewer instruction", "operator instruction", or "operator command" can be obtained from explicit commands entered by the viewer or may be implicitly obtained or derived based on some other user action, such as making an equipment setting, for example. With respect to entries entered on an operator interface, such as an interface using a display monitor and keyboard, for example, the terms "command" and "instruction" may be used interchangeably to refer to an operator entry.

In the context of the present disclosure, a single projected line of light is considered a "one dimensional" pattern, since the line has an almost negligible width, such as when projected from a line laser, and has a length that is its predominant dimension. Two or more of such lines projected side by side, either simultaneously or in a scanned arrangement, provide a two-dimensional pattern. In exemplary embodiments, lines of light can be linear, curved or three- dimensional.

The terms "3-D model", "point cloud", "3-D surface", and "mesh" may be used synonymously in the context of the present disclosure. The dense point cloud is formed using techniques familiar to those skilled in the volume imaging arts for forming a point cloud and relates generally to methods that identify, from the point cloud, vertex points corresponding to surface features. The dense point cloud is thus generated using the reconstructed contour data from one or more reflectance images. Dense point cloud information serves as the basis for a polygon model at high density for the teeth and gum surface.

Apparatus and/or method embodiments according to the application aim at virtual definition of a base of a dental virtual model for facilitating a repeatable, accurate and rapid orientation of a virtual dental model to the base. Certain exemplary embodiments can provide automatic formation of digital model bases from a plurality of prescribed dimensions and bounding dimensions of a 3-D digital model of the patient's dentition. One exemplary use of virtual dental models is for virtual storage purposes in orthodontic treatment.

A virtual teeth model is received, provided, or reconstructed (e.g., from a laser scanning or intra oral camera scanning of a plaster, a negative impression (e.g., alginate or silicon) of a patient's dentition or directly the patient's mouth). The virtual teeth model can also be obtained from an x-ray scan of a plaster model or a negative impression of the patient's teeth. Thus, the virtual teeth models can be obtained using intraoral or extraoral scanning

devices/methods. The virtual teeth models can be positioned on a virtual base or virtual base model. One exemplary use of this virtual teeth model is for virtual storage purposes in orthodontic treatment.

In one embodiment, a virtual teeth model of a patient's dentition can be obtained from an intraoral scanner.

Figure 1 is a schematic diagram showing an imaging apparatus 70 for projecting and imaging using structured light patterns 46. Imaging apparatus 70 uses a handheld camera 24 for image acquisition according to an embodiment of the present disclosure. A control logic processor 80, or other type of computer that may be part of camera 24 controls the operation of an illumination array 10 that generates the structured light and controls operation of an imaging sensor array 30. Image data from surface 20, such as from a tooth 22, is obtained from imaging sensor array 30 and stored in a memory 72. Control logic processor 80, in signal communication with camera 24 components that acquire the image, can process the received image data and stores the mapping in memory 72. The resulting image from memory 72 is then optionally rendered and displayed on a display 74. Memory 72 may also include a display buffer for temporarily storing display 74 image content. Control logic processor 80 can be part of a standalone computer or workstation with a separate display and user input functions. Control logic processor 80 can be remotely located relative to a camera 24.

In fringe projection imaging of a surface, a pattern of lines is projected from illumination array 10 toward the surface of an object from a given angle. The projected pattern from the surface is then viewed from another angle as a contour image, taking advantage of triangulation in order to analyze surface information based on the appearance of contour lines. Phase shifting, in which the projected pattern is incrementally shifted spatially for obtaining additional measurements at the new locations, is typically applied as part of fringe projection imaging, used in order to complete the contour mapping of the surface and to increase overall resolution in the contour image.

The schematic diagram of Figure 2 shows, with the example of a single line of light L, how patterned light is used for obtaining surface contour information using a handheld camera or other portable imaging device. A mapping is obtained as an illumination array 10 directs a pattern of light onto a surface 20 and a corresponding image of a line L' is formed on an imaging sensor array 30. Each pixel 32 on imaging sensor array 30 maps to a corresponding pixel 12 on illumination array 10 according to modulation by surface 20. Shifts in pixel position, as represented in Figure 2, yield useful information about the contour of surface 20. It can be appreciated that the basic pattern shown in Figure 2 can be implemented in a number of ways, using a variety of illumination sources and sequences and using one or more different types of sensor arrays 30. Illumination array 10 can utilize any of a number of types of arrays used for light modulation, such as a liquid crystal array or digital micromirror array, such as that provided using the Digital Light Processor or DLP device from Texas Instruments, Dallas, TX. This type of spatial light modulator is used in the illumination path to change the light pattern as needed for the mapping sequence.

By projecting and capturing images that show structured light patterns that duplicate the arrangement shown in Figures 1 and 2 multiple times, the image of the contour line on the camera simultaneously locates a number of surface points of the imaged object. This can speed the process of gathering many sample points, while the plane of light (and usually also the receiving camera) is laterally moved in order to "paint" some or all of the exterior surface of the object with the plane of light.

Figure 3 shows surface imaging using a pattern with multiple lines of light. Incremental shifting of the line pattern and other techniques help to compensate for inaccuracies and confusion that can result from abrupt transitions along the surface, whereby it can be difficult to positively identify the segments that correspond to each projected line. In Figure 3, for example, it can be difficult to determine whether line segment 16 is from the same line of illumination as line segment 18 or adjacent line segment 19.

By knowing the instantaneous position of the camera and the instantaneous position of the line of light within an object-relative coordinate system when the image was acquired, a computer and software can use triangulation methods to compute the coordinates of numerous illuminated surface points. As the plane is moved to intersect eventually with some or all of the surface of the object, the coordinates of an increasing number of points are accumulated. As a result of this image acquisition, a point cloud of vertex points or vertices can be identified and used to represent the extent of a surface within a volume. By way of example, Figure 4 shows a dense point cloud 50 generated from a structured light imaging apparatus, CS 3500 3-D camera made by

Carestream Heath, Inc., Rochester NY, USA, using results from patterned illumination such as that shown in Figure 3. The point cloud 50 models physical location of sampled points on tooth surfaces and other intraoral surfaces or, more generally, of surfaces of a real-world object. Variable resolution can be obtained. The example of Figure 4 shows an exemplary 100 micron resolution. The points in the point cloud represent actual, measured points on the three dimensional surface of an object.

The surface structure can be approximated from the point cloud representation by forming a polygon mesh, in which adjacent vertices are connected by line segments. For a vertex, its adjacent vertices are those vertices closest to the vertex in terms of Euclidean distance.

By way of example, Figure 5 shows a 3-D polygon mesh model 60

(e.g., of a patient's dentition) in the simple form of a triangular mesh. A triangular mesh forms a basic mesh structure that can be generated from a point cloud and used as a digital model to represent a 3-D object by its approximate surface shape, in the form of triangular plane segments sharing adjacent boundaries. Methods/apparatus for forming a polygon mesh model, such as a triangular mesh or more complex mesh structure, are well known to those skilled in the contour imaging arts. The logic flow diagram of Figure 6 shows an exemplary method embodiment for virtual teeth base generation (e.g., orthodontic base models) according to the present disclosure. Starting from the reconstructed 3D virtual teeth model, certain exemplary method and/or apparatus embodiments provide a virtual treatment of a base of this virtual teeth model (virtual reconstructed teeth model). The virtual teeth model can be or is required to be shown on a virtual orthodontic base, with the consideration or respect of the regulatory requirements (for example, the American Board of Orthodontics (ABO) policy). Certain exemplary virtual base embodiments herein are adapted to at least the size of the virtual teeth model.

As shown in Figure 6, in a virtual base requirements step S600, a plurality of virtual base requirements of rules can be obtained. Some exemplary virtual base embodiments herein respect at least the virtual orthodontic base regulatory requirements of the American Board of Orthodontics (ABO). For example, certain exemplary virtual base embodiments herein respect at least the following rules:

❖ The length of the oblique faces on the rear corner of the base (b) is 13.0mm

❖ The indicated angles are 65.0° (e.g., the direction of all the faces are imposed by the regulatory agencies)

❖ The minimal height of the base is 13.0mm

❖ The distance between a maxillary base and a mandibular base is

comprised between 60.0 and 70.0 mm.

Embodiments of the application are not intended to be so limited, however, since in alternative embodiments, additional rules, different rules can be followed by such an exemplary dental virtual base embodiment.

Figure 7a is a diagram that shows a perspective view of an exemplary virtual dental base model embodiment (e.g., maxillary arch and/or mandibular arch) according to the application. As shown in Figure 7a, an exemplary virtual dental base model embodiment 700 can include upper and/or lower portions (e.g., maxillary arch and/or mandibular arch) according to the application. A first distance 707 between a lower portion 750 and an upper portions 760 is preferably between a first prescribed range so a height of the model 700 is between a prescribed height range 708.

Figure 7b is a diagram that shows a top view of an exemplary footprint for a virtual dental base model embodiment (e.g., maxillary arch) according to the application. Figure 7c is a diagram that shows a cross-sectional view of an exemplary virtual dental base model embodiment in Figure 7b according to the application. As shown in Figure 7b, an oblique face 702 of a virtual base 750 has a prescribed length and selected angles 704 have a prescribed arc. As shown in Figure 7c, upper portion 750 (or lower portion 760) of the virtual base 700 is higher than a prescribed height 706 (e.g., minimum height).

Figure 7c is a diagram that shows a perspective view of an exemplary virtual dental base model embodiment (e.g., maxillary arch and/or mandibular arch) according to the application. As shown in Figure 7c, a distance 708 between the lower and/or upper portions 750, 760 is between a prescribed range.

Figure 7d is a diagram that shows a top view of an exemplary footprint for a virtual dental base model embodiment (e.g., mandibular arch) according to the application. Figure 7e is a diagram that shows a cross-sectional view of an exemplary virtual dental base model embodiment (e.g., mandibular arch) according to the application. As shown in Figure 7d, an oblique face 702 of a virtual base 700 has a prescribed length and selected angles 704 have a prescribed arc. As shown in Figure 7e, lower and/or upper portions 750, 760 of the virtual base 700 are higher than a prescribed height 706. In the exemplary embodiment of Figure 7d, the mandibular base 760 can have the same rules as the maxillary base, but can have a curved shaped (e.g., a circle shape) in the front. In one embodiment, angle 704 can be 65 degrees, a prescribed height threshold 706 can be 13 mm, and length 702 can be 13mm.

The dental frenum (or frenulum), is a small fold of skin between the lip and the gum. In certain exemplary embodiments, the frenulums can be represented as small peaks on the surface of the virtual base. As shown in Figures 7a-7e, on the right and left sides of the base, we can see the dental frenulum 720. In one embodiment, angle 704 can be 65 degrees and length 702 can be 13mm. An obtain teeth 3-D model step S610 then inputs 3-D digital model of the patient's dentition. As described herein, the 3-D digital model of the patient's dentition can be obtained as known to one skilled in the art (e.g., using extraoral or intraoral techniques). A determine bounding dimensions of the 3-D digital model of the patient's dentition step S620 can determine a 3-D enclosure surrounding the 3-D digital model of the patient's dentition. As shown in Figure

8, a 3-D bounding box 810 is determined to surround a 3-D digital model 820 of the patient's dentition. In Figure 8, the bounding box 810 is illustrated as a 2D rectangle with dimensions to surround both the upper and lower dental arch models. In certain exemplary embodiments, the bounding box 810 can surround the upper and lower dental arch models that are in an occlusal position or relationship.

A transformation step S630 can then preferably automatically generate virtual base model. In certain exemplary embodiments, the

transformation step S630 can include global and local transformations based on bounding dimensions and local criteria. The dimensions of the maxillary and mandibular bases (e.g., height and length of the faces as well as the total height of both models 750, 760 put together in occlusion (and optionally a position of the frenulums), are preferably automatically adapted to the size and shape of the maxillary and mandibular reconstructed virtual teeth models. As shown in Figure

9, from the 3D bounding box 810 of the global teeth model (e.g., maxillary and mandibular), an initial desired or optimal size can be determined for the virtual base(s) 750, 760. The initial virtual base size can then undergo global transformations adjusting the virtual base as a whole.

Global transformations provided by certain exemplary embodiments include (i) translation of the faces to be placed at the optimal distance (e.g., required height) and/or (ii) global scaling from front to bottom and from right to left. Figure 10 is a diagram that illustrates exemplary global transformations applied to an initially sized virtual base model. In one exemplary embodiment, an initial virtual base(s) can be automatically deformed to have the selected size that corresponds to current aesthetic criteria and/or to a current practice for the definition of the model base of the real plasters of patient teeth. In alternate exemplary embodiments, some manual input can also be contemplated or used. In step S730, the globally transformed virtual base model can then undergo local transformations. In one embodiment, exemplary local transformations can be performed section by section (e.g., rear sides, middle sides and front sides. Local transformations provided by certain exemplary embodiments include local movement of the main or primary points of the virtual base (e.g., corner, front points, frenulums, etc). Figure 1 1 is a diagram that illustrates exemplary local transformations applied to a globally transformed virtual base model. As shown in Figure 1 1 , for each section, the height of the globally transformed base can then be locally adapted, by changing the local height of the base along the illustrated arrows.

A display step S630 can then preferably display a final virtual base model assembled with the virtual teeth model. The final virtual base models according to exemplary method and/or apparatus embodiments of the application still maintain the prescribed requirements (e.g., respect the norms of the ABO defined requirements. The final virtual base model can alternatively be stored or transmitted for use. Exemplary definition of the parametric virtual base:

Exemplary parameters for a virtual base model determination embodiment are illustrated in Figure 12. In one exemplary process embodiment, a virtual dynamic base 1200 can be determined, entirely parametrically, with the following algorithmic relationship shown in equation ( 1 ):

❖ d = [ (a + a')tanp -h]/q

❖ c=(a+a')/q +h*[l/sinp - l/(q*tanp)]-b Equation ( 1 ) where:

❖ q=sinptanp - cosp

❖ a' = b*sqrt(2-2cosp»

If a scale is applied to the virtual dynamic base 1200 (e.g., scale value in X = S a and scale value in Y = S h ), then the various length change as following equation (2): ❖ Aa = (S a - 1.0) * a

❖ Ab = 0.0

❖ A h = (S„ - 1.0) * h

❖ A d = Aa * tanp /q - A„ /q

❖ A c = >/[ A d * A d + A„ * A h -2* A d * A h * οοββ] Equation (2)

Preferably, the virtual base model is not a uniform mesh. In certain exemplary embodiments, the virtual base model 1200 contains some points and triangles. Figure 13 is a diagram that illustrates an exemplary polygonal (e.g., triangular) mesh forming the virtual base model 1200.

For each point of the base P (P x ,P y , P z ), as shown in Figure 13, a position can be defined, which depends only of a, h, c, d, b and β, and so only depends on a, h, b and β (because c and d depends only on a, h, b and β).

When the scale is applied, then the new position of the point P is the following equation (3):

P' = P + w a * A a X +w c * A c V c + w d * A d *V d Equation (3)

If (- (b+ c) * 5ϊηβ < Py < -b * 5Ϊηβ )

w d = 0.0

w c = 1.0

The vector V d and V c are different if Px is positive or negative. Accordingly, a deformation of each point on a virtual base model according to the deformation

(e.g., global or local) applied on the base can be determined.

Figure 14 is a diagram that illustrates local movements along an exemplary virtual base model embodiment. As shown in Figure 14, for the local movements (e.g., local z scale and local xy displacement), xy local displacement can be done only along the segment where the point is positioned. For example, if

P belongs to the front right edge, then it can be moved only along the vector V d .

For the local z scale: a point belonging to a vertical face can be moved according to its distance to the associated edge and bottom points. Certain exemplary method and/or apparatus embodiments according to the application can provide virtual definition of the base of a dental virtual model. Exemplary embodiments according to the application can include various features described herein (individually or in combination). Exemplary method embodiments shown in Figure 6 can be can be performed by standalone computers, workstations or the like, or the same or control logic processors integral to dental imaging apparatus/ dental x-ray systems; however, method embodiments of Fig. 6 is not intended to be limited thereby.

In certain exemplary embodiments, parameters for a virtual base model determination such as determined by equations ( l )-(3) can be used to implement step S630 in whole or in part; however, method embodiments of Fig. 6 is not intended to be limited thereby.

Certain exemplary embodiments can provide formation of individualized digital model bases for each patient from a plurality of prescribed dimensions and bounding dimensions of a 3-D digital model of said each patient's dentition. Certain exemplary method and/or apparatus embodiments can provide formation of personalized digital model bases for each patient (e.g., dentition models) that remain compliant with ABO regulations and/or promulgated regulatory agency requirements.

Consistent with one embodiment of the present disclosure, the present disclosure utilizes a computer program with stored instructions that perform on image data accessed from an electronic memory. As can be appreciated by those skilled in the image processing arts, a computer program of an embodiment of the application can be utilized by a suitable, general-purpose computer system, such as a personal computer or workstation. However, many other types of computer systems can be used to execute the computer program of the application, including networked processors. The computer program for performing the method of the application may be stored in a computer readable storage medium. This medium may comprise, for example; magnetic storage media such as a magnetic disk such as a hard drive or removable device or magnetic tape; optical storage media such as an optical disc, optical tape, or machine readable bar code; solid state electronic storage devices such as random access memory (RAM), or read only memory (ROM); or any other physical device or medium employed to store a computer program. The computer program for performing the method of the application may also be stored on computer readable storage medium that is connected to the image processor by way of the internet or other communication medium. Those skilled in the art will readily recognize that the equivalent of such a computer program product may also be constructed in hardware.

It should be noted that the term "memory", equivalent to

"computer-accessible memory" in the context of the application, can refer to any type of temporary or more enduring data storage workspace used for storing and operating upon image data and accessible to a computer system, including a database, for example. The memory could be non-volatile, using, for example, a long-term storage medium such as magnetic or optical storage. Alternately, the memory could be of a more volatile nature, using an electronic circuit, such as random-access memory (RAM) that is used as a temporary buffer or workspace by a microprocessor or other control logic processor device. Displaying an image requires memory storage. Display data, for example, is typically stored in a temporary storage buffer that is directly associated with a display device and is periodically refreshed as needed in order to provide displayed data. This temporary storage buffer can also be considered to be a memory, as the term is used in the application. Memory is also used as the data workspace for executing and storing intermediate and final results of calculations and other processing. Computer-accessible memory can be volatile, non-volatile, or a hybrid combination of volatile and non-volatile types.

It will be understood that the computer program product of the application may make use of various image manipulation algorithms and processes that are well known. It will be further understood that the computer program product embodiment of the application may embody algorithms and processes not specifically shown or described herein that are useful for implementation. Such algorithms and processes may include conventional utilities that are within the ordinary skill of the image processing arts. Additional aspects of such algorithms and systems, and hardware and/or software for producing and otherwise processing the images or co-operating with the computer program product of the present disclosure, are not specifically shown or described herein and may be selected from such algorithms, systems, hardware, components and elements known in the art.

While the invention has been illustrated with respect to one or more implementations, alterations and/or modifications can be made to the illustrated examples without departing from the spirit and scope of the appended claims. In addition, while a particular feature of the invention can have been disclosed with respect to one of several implementations, such feature can be combined with one or more other features of the other implementations as can be desired and advantageous for any given or particular function. The term "at least one of is used to mean one or more of the listed items can be selected. The term "about" indicates that the value listed can be somewhat altered, as long as the alteration does not result in nonconformance of the process or structure to the illustrated embodiment. Finally, "exemplary" indicates the description is used as an example, rather than implying that it is an ideal. Other embodiments of the invention will be apparent to those skilled in the art from consideration of the specification and practice of the invention disclosed herein. The presently disclosed embodiments are therefore considered in all respects to be illustrative and not restrictive. The scope of the invention is indicated by the appended claims, and all changes that come within the meaning and range of equivalents thereof are intended to be embraced therein.