WO2018132321A1 - Diagnostic methods for the classifiers and the defects captured by optical tools - Google Patents
Diagnostic methods for the classifiers and the defects captured by optical tools Download PDFInfo
- Publication number
- WO2018132321A1 WO2018132321A1 PCT/US2018/012684 US2018012684W WO2018132321A1 WO 2018132321 A1 WO2018132321 A1 WO 2018132321A1 US 2018012684 W US2018012684 W US 2018012684W WO 2018132321 A1 WO2018132321 A1 WO 2018132321A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- wafer
- inspection results
- interest
- processor
- inspection
- Prior art date
Links
Classifications
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L22/00—Testing or measuring during manufacture or treatment; Reliability measurements, i.e. testing of parts without further processing to modify the parts as such; Structural arrangements therefor
- H01L22/30—Structural arrangements specially adapted for testing or measuring during manufacture or treatment, or specially adapted for reliability measurements
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B19/00—Programme-control systems
- G05B19/02—Programme-control systems electric
- G05B19/418—Total factory control, i.e. centrally controlling a plurality of machines, e.g. direct or distributed numerical control [DNC], flexible manufacturing systems [FMS], integrated manufacturing systems [IMS] or computer integrated manufacturing [CIM]
- G05B19/41875—Total factory control, i.e. centrally controlling a plurality of machines, e.g. direct or distributed numerical control [DNC], flexible manufacturing systems [FMS], integrated manufacturing systems [IMS] or computer integrated manufacturing [CIM] characterised by quality surveillance of production
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N7/00—Computing arrangements based on specific mathematical models
- G06N7/01—Probabilistic graphical models, e.g. probabilistic networks
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L21/00—Processes or apparatus adapted for the manufacture or treatment of semiconductor or solid state devices or of parts thereof
- H01L21/67—Apparatus specially adapted for handling semiconductor or electric solid state devices during manufacture or treatment thereof; Apparatus specially adapted for handling wafers during manufacture or treatment of semiconductor or electric solid state devices or components ; Apparatus not specifically provided for elsewhere
- H01L21/67005—Apparatus not specifically provided for elsewhere
- H01L21/67242—Apparatus for monitoring, sorting or marking
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L22/00—Testing or measuring during manufacture or treatment; Reliability measurements, i.e. testing of parts without further processing to modify the parts as such; Structural arrangements therefor
- H01L22/20—Sequence of activities consisting of a plurality of measurements, corrections, marking or sorting steps
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L22/00—Testing or measuring during manufacture or treatment; Reliability measurements, i.e. testing of parts without further processing to modify the parts as such; Structural arrangements therefor
- H01L22/20—Sequence of activities consisting of a plurality of measurements, corrections, marking or sorting steps
- H01L22/24—Optical enhancement of defects or not directly visible states, e.g. selective electrolytic deposition, bubbles in liquids, light emission, colour change
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B2219/00—Program-control systems
- G05B2219/30—Nc systems
- G05B2219/37—Measurements
- G05B2219/37224—Inspect wafer
-
- H—ELECTRICITY
- H01—ELECTRIC ELEMENTS
- H01L—SEMICONDUCTOR DEVICES NOT COVERED BY CLASS H10
- H01L22/00—Testing or measuring during manufacture or treatment; Reliability measurements, i.e. testing of parts without further processing to modify the parts as such; Structural arrangements therefor
- H01L22/10—Measuring as part of the manufacturing process
- H01L22/12—Measuring as part of the manufacturing process for structural parameters, e.g. thickness, line width, refractive index, temperature, warp, bond strength, defects, optical inspection, electrical measurement of structural dimensions, metallurgic measurement of diffusions
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02P—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
- Y02P90/00—Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
- Y02P90/02—Total factory control, e.g. smart factories, flexible manufacturing systems [FMS] or integrated manufacturing systems [IMS]
Definitions
- the present disclosure relates to defect detection.
- Fabricating semiconductor devices typically includes processing a semiconductor wafer using a large number of fabrication processes to form various features and multiple levels of the semiconductor devices.
- lithography is a semiconductor fabrication process that involves transferring a pattern from a reticle to a photoresist arranged on a semiconductor wafer.
- Additional examples of semiconductor fabrication processes include, but are not limited to, chemical-mechanical polishing (CMP), etch, deposition, and ion implantation.
- CMP chemical-mechanical polishing
- etch etch
- deposition deposition
- ion implantation ion implantation
- Algorithms can be used to detect defects on a wafer.
- machine learning algorithms are used to create defect classifiers and nuisance filters, the algorithms tend to be considered as black box solutions that are not tuned or diagnosed.
- Assessment of an inspection recipe typically waited until observing a new set of labeled data received for the assessment or, alternatively, not using some part of the labeled data and reserving it for the validation. Both of these techniques wasted resources.
- the overall performance assessment may be done based on the quality of the data used to train the classifier and the classifier's ability to learn and extract the information from the data. If the quality of the data is poor and the real defects and nuisance do not have a clear separation boundary, then any classifier will likely fail.
- the performance of each recipe is evaluated using two measures: discriminability and reliability.
- discriminability measures There are many discriminability measures.
- One is the confusion matrix of the training data, which consists of a set of conditional error rates. From these conditional error rates, the cap rate and nuisance rate may be important to semiconductor manufacturers.
- Cap rate is the ratio of the number of defects of interest (DOI) that are classified correctly to the total number of DOI in the wafer.
- the nuisance rate is the number of nuisance that are classified as DOI to the total number of defects that are classified as DOI.
- a higher cap rate and a lower nuisance rate mean a better recipe. However, nuisance rate and cap rate could only previously be evaluated for the training data set that included the actual labels of data.
- Reliability is a measure that shows how certain the classifier is about the decision it makes. It is a function of the estimation of posteriors done by the classifier. Previously, the classifier reliability was evaluated through the confidence calculation for each defect. [0009] Although discriminability and reliability can be important measures, discriminability and reliability can cover up the reality if the underlying distributions of DOI and nuisance have certain characteristics. This can be referred to as a shadowing effect.
- the confusion matrix of the training set typically is not sufficient to understand the behavior of the recipe on the entire wafer. If the defects in the training set have been selected in certain way (which normally is done in order to reduce the number of defects for scanning electron microscope (SEM) review and manual classification), then the confusion matrix of the training set is biased toward those defects and is not be a good estimator of the classifier performance over the entire wafer.
- SEM scanning electron microscope
- Previous solutions retrain a binary classifier (e.g., nuisance vs. DOI) based on manual classification obtained during process monitoring (production sampling). These previous solutions used the updated classifier to create new DOI/nuisance separation on the subsequent wafer and used the new bins to generate production sample, which in turn was used to tune the next classifier. 50% of the previous solutions' sample is random sampling from the DOI bin of the latest classifier, and the other 50% is random sampling from the entire population. The two samples are used to compare the statistical process control (SPC) of the two inspections, and the second sample also provides "sub-threshold" defects to be used for retraining of the classifier.
- SPC statistical process control
- a system for detecting defects of interest in a plurality of wafers comprises a central storage media configured to store a plurality of classified inspection results and an initial defect classifier; a wafer inspection tool; an image data acquisition system; and a processor in electronic communication with the central storage media, the wafer inspection tool, and the image data acquisition system.
- the processor is configured to execute the instructions of: an inspection engine; a sampling engine, and a tuning engine.
- the inspection engine instructs the processor to receive inspection results of a first wafer from the wafer inspection tool.
- the sampling engine instructs the processor to: retrieve the initial defect classifier from the central storage media; filter the inspection results based on the initial defect classifier; review locations of interest on the first wafer from the image data acquisition system based on the filtered inspection results; classify the filtered inspection results based on the initial defect classifier; store the classified filtered inspection results in the central storage media; and identify defects of interest in the first wafer based on the classified filtered inspection results.
- the tuning engine instructs the processor to update the initial defect classifier based on the stored classified filtered inspection results in the central storage media. For each remaining wafer, the inspection engine instructs the processor to receive inspection results of a next wafer from the wafer inspection tool.
- the sampling engine instructs the processor to: filter the inspection results of the next wafer based on the initial defect classifier; review locations of interest on the next wafer, using the image data acquisition system based on the filtered inspection results of the next wafer and historical analysis sampling; classify the filtered inspection results of the next wafer based on the reviewed locations of interest on the next wafer; store the classified filtered inspection results for the next wafer in the central storage media; update the defect classifier, using the processor, based on the stored classified filtered inspection results for the next wafer in the central storage media; and identify defects of interest in the next wafer based on the classified filtered inspection results for the next wafer.
- the tuning engine can instruct the processor to update the defect classifier, using the processor, based on the stored classified filtered inspection results for the next wafer in the central storage media.
- the sampling engine can instruct the processor to perform the filtering step based on the updated defect classifier.
- the image data acquisition system can be an SEM review tool.
- the wafer inspection tool can perform a hot scan to capture inspection results.
- the wafer inspection tool can be a broadband plasma inspection tool.
- the defect classifier can send defect of interest data and nuisance data for retraining of the defect classifier.
- the step of identifying defects of interest can comprise: sampling near a classification boundary of a most recent defect classifier; obtaining information about classifier stability based on fluctuations in the defect classifier; observing a movement in the classification boundary; and identifying the defects of interest based on the predicted movement in the classification boundary.
- the inspection results or reviewed locations of interest can be stored in the central storage media.
- a method for identifying defects of interest in a plurality of wafers comprises receiving, at a processor, inspection results of a first wafer from a wafer inspection tool. Using the processor, the inspection results are filtered based on an initial defect classifier. Locations of interest on the first wafer are reviewed, using an image data acquisition system, based on the filtered inspection results. The filtered inspection results are classified, using the processor, based on the reviewed locations of interest on the first wafer. The classified filtered inspection results are stored in a central storage media. Defects of interest in the first wafer are identified based on the classified filtered inspection results.
- the method comprises receiving, at the processor, inspection results of a next wafer from the wafer inspection tool.
- the inspection results are filtered based on the initial defect classifier. Locations of interest on the next wafer are reviewed, using the image data acquisition system, based on the filtered inspection results of the next wafer and historical analysis sampling.
- the filtered inspection results of the next wafer are classified, using the processor, based on the reviewed locations of interest on the next wafer.
- the classified filtered inspection results for the next wafer are stored in the central storage media.
- the defect classifier is updated, using the processor, based on the stored classified filtered inspection results for the next wafer in the central storage media. Defects of interest in the next wafer are identified based on the classified filtered inspection results for the next wafer.
- the image data acquisition system can be an SEM review tool.
- the wafer inspection tool can perform a hot scan to capture inspection results.
- the wafer inspection tool can be a broadband plasma inspection tool.
- the defect classifier can send defect of interest data and nuisance data for retraining of the defect classifier.
- the step of identifying defects of interest can comprise: sampling near a classification boundary of a most recent defect classifier; obtaining information about classifier stability based on fluctuations in the defect classifier; observing a movement in the classification boundary; and identifying the defects of interest based on the predicted movement in the classification boundary.
- the method can comprise updating the defect classifier, using the processor, based on the stored classified filtered inspection results for the next wafer in the central storage media.
- the filtering step can be performed based on the updated defect classifier.
- the inspection results or reviewed locations of interest can be stored in the central storage media.
- the step of updating the defect classifier based on the stored classified filtered inspection results in the central storage media can comprise: estimating a cap rate based on a calculated training confusion matrix and estimating a nuisance rate based on the defect classifier in the central storage media, the classified filtered inspection results for the next wafer, and the estimated cap rate.
- the calculated training confusion matrix is based on the stored classified filtered inspection results for the next wafer in the central storage media.
- the filtered inspection results can have at least two thresholds associated with the filtered inspection results. A first of the at least two thresholds is for an inspection used for monitoring processes and defects. A second of the at least two thresholds is less than the first threshold and is configured to capture sub-threshold defects during inspection.
- Figure 1 includes flowcharts (a) and (b) of previous techniques
- Figure 2 includes charts (a), (b), and (c) of a cap rate versus cutline curve, a nuisance rate versus cutline curve, and a cap rate versus nuisance rate curve, respectively;
- Figure 3 includes distributions (a), (b), and (c);
- Figure 4 includes distributions (a) and (b);
- FIG. 5 is a flowchart of an embodiment of a shadow detect algorithm in accordance with the present disclosure
- Figure 6 includes charts of accuracy versus number of defects in the pool for an ordinary wafer (a) and a shadowed wafer (b);
- Figure 7 is a flowchart of an embodiment of a diagnostic model in accordance with the present disclosure.
- FIG. 8 is a flowchart of an embodiment in accordance with the present disclosure.
- Figure 9 is a block diagram of a system in accordance with the present disclosure.
- Figure 10 is a diagram of a dynamic classifier with dynamic sampling and stability analysis in accordance with the present disclosure.
- Figure 11 is a diagram of a static classifier with dynamic sampling and stability analysis in accordance with the present disclosure.
- the embodiments disclosed herein address new systems and methods for dealing with process and wafer instability in the early stages of an integrated circuit manufacturing process.
- One embodiment of the present disclosure is based on the idea of producing a small sample on productions lots in addition to the production sampling, aggregating the sample over several wafers to build an up-to-date classifier, and using classifier to produce new updated samples on the next wafer.
- the embodiments disclosed herein may be particularly advantageous over existing methods for at least the following reasons.
- the presently disclosed systems and methods utilize a supplemental (augmented) sample which is generated using the latest known process conditions and is well-suited for returning a superior classifier.
- the latest known process conditions and defects are far more useful for this purpose than the random samples currently used.
- the supplemental sample can be automatically tuned to match those process conditions.
- disclosed systems and methods allow for more relevant broadband plasma inspection with more stable nuisance rate and DOI capture rates.
- the disclosed systems and methods allow for faster discovery of newly appearing defects occurring during the manufacturing process, and allow for an analysis of the stability of the manufacturing process.
- One embodiment relies solely on data from the central storage media and the systems and methods leverage the manual classification of defects in the classifier performance on the rest of the inspection. These embodiments cause the classifier to be one wafer behind what is currently being inspected.
- Another embodiment adds the ability to update the classifier on the current wafer by performing sampling on the wafer defect review tool and then producing the supplemental sample the central storage media.
- One advantage of this embodiments is that the latest wafer condition is also included in the classifier.
- the cap rate and the nuisance rate can be estimated for the data where the actual labels are not available.
- the expected values for the cap rate and the nuisance rate can be provided.
- the techniques show that all the estimations of cap rate, nuisance rate, posteriors, and confidence are accurate or that the data has the shadowed distribution.
- Data produced by algorithms can provide diagnostics information that is not available with manually created classifiers such as inline defect organizer (iDO).
- iDO inline defect organizer
- a recipe can be assessed.
- ROC receiver operating curve
- DOI cap rate true positive rate
- nuisance rate which is not false positive rate
- Two outputs from the classifier can be used to build the diagnostic tools.
- decisions which are the classification results provided by the classifier, can be used.
- posteriors for each defect can be used. There are different ways that a classifier might find posteriors. Distance from each class centroid or probabilistic measures of accuracy are two examples.
- the ratio of the number of the DOIs that are correctly classified to the total number of DOIs in the training set can be used. This can be applied to the test set to find the estimation of the number of the DOI potentially missed in the test data.
- Snm is the set of all defects that originally belongs to class m and classified as class n.
- SDD is the set of defects that are classified as DOI and are actually DOI.
- SND is the set of defects that are classified as nuisance and are actually DOI.
- SDN is the set of defects that are classified as DOI and are actually nuisance.
- SNN is the set of defects that are classified as nuisance and are also actually nuisance.
- Equation 1
- denotes the size (cardinality) of set S.
- denotes the size (cardinality) of set S.
- SD is the set of defects that are classified as DOI.
- SN is the set of defects that are classified as nuisance. Assuming the posterior for the nuisance class associated to defect i is pi, the nuisance rate would be calculated as shown in Equation 2.
- SD is the set of defects that are classified as DOI.
- pi is the posterior probability for the nuisance class associated to defect i.
- denotes the size of set SD.
- the cap rate can be increased with higher nuisance. This can be done by, for example, moving the cut lines in the confidence histogram and changing the class codes of the defects with lower confidence.
- the cap rate and nuisance rate can be evaluated for all the possible values of the cutline. Then the three plots can be shown, three examples of which are shown in Figure 2.
- Chart (a) in Figure 2 shows cap rate versus cutline value.
- Chart (b) in Figure 2 shows nuisance rate versus cutline value.
- Chart (c) in Figure 2 shows ROC.
- An ROC can be a useful representation of the performance of a classifier on a given data set. The user can find what is going to be the nuisance rate for a desired capture rate, and vice versa. With these curves, a user can decide if the values of cutline worthwhile.
- well-separated distributions may be the ones with short overlap as shown in Figure 3(a), (b), and (c). Data can be well-separated as if a clear boundary can be draw between the two distributions as shown in Figure 3 (a) and (b). Distributions can be well- separated and have multiple regions in the space and be separated using multiple boundaries as shown in Figure 3(c). [0046] Most classifiers can learn this situation. In this scenario, the performance of the classifier is ordinary. Such Probability Density Functions (PDFs) are the ones that normally appears in the wafer, but this is not always the case. A big part of one distribution may have been shadowed.
- PDFs Probability Density Functions
- Shadowing effect is a situation when a large part of one class distribution is under the PDF of another class. This situation can happen as mistakes during manual or automated labeling or as a result of not having good attributes to distinguish the shadowed part from the other class. Charts (a) and (b) in Figure 4 are two examples of this situation.
- the detection of the first case ((a) in Figure 4) is relatively easy because, just by observing the training confusion matrix, it can be determined that the accuracy for one class is poor. Detecting the second case ((b) in Figure 4) is more difficult. This situation can mislead a user about the data on the wafer where big portion of one class will not be detected, no matter what kind of the classifier is used. The misclassification here is not due to the poor performance of the classifier, but may be due to the poor quality of features or labeling.
- a classifier can be trained with the training set. Then, the training can be sorted to set ascendingly from the confidence values the defects obtained from the classifier. An empty pool can be created and defects can be added one by one from lowest confidence to highest confidence to the pool. After adding each defect, the confusion matrix of the defects in the pool can be calculated and the accuracies of the classes and the number of defects in the pool can be saved. Accuracy for each class can be defined as the number of correctly classified defects of that class to the total number of defects from that class. After using all the defects in the training set, the accuracies versus the number of defects in the pool can be compared. An example of this algorithm is shown in Figure 5.
- (a) of Figure 6 is from a wafer without shadowed DOI and both DOI and nuisance accuracies improve with the number of defects.
- the plot (b) in Figure 6 shows a wafer which a DOI class observes a shadowing effect.
- the DOI bin does not improve with the number of defects. It indicates that high confidence defect are added, but these are being classified incorrectly, which is an indication of the shadowing effect.
- Figure 7 shows a flow chart for the algorithm of estimating nuisance rate and capture rate and detection of the shadow effect.
- the training set is used to create the classifier.
- the classifier is applied to the defects in the test set.
- the classifier is used to evaluate the confidence and posteriors for all the defects (both in the training set and test set).
- the estimation of the nuisance rate is done using the posteriors.
- the estimation of the capture rate is done using the confusion matrix obtained from the training set. Finally, a check is done to find out if the data is under the shadow effect or not. If it is not, then the estimates are trustable.
- Figure 8 is a flowchart of a method 100 for identifying defects of interest in a plurality of wafers.
- inspection results of a first wafer are received, such as at a processor, from a wafer inspection tool, which may be a BBP tool or another inspection device.
- the inspection results are filtered based on an initial defect classifier, such as using the processor.
- locations of interest on the first wafer are reviewed based on the filtered inspection results, such as using an image data acquisition system.
- the image data acquisition system may be an SEM review tool or another measurement, inspection, or metrology tool.
- the filtered inspection results are classified, such as using the processor, based on the reviewed locations of interest on the first wafer.
- the classified inspection results are stored in a central storage media.
- defects of interest are identified based on the classified filtered inspection results, such as using the processor. Filtered inspection results may be kept separate, such as for each wafer that is sampled.
- inspection results of the next wafer are received, such as at the processes, from the wafer inspection tool at 107.
- the inspection results are filtered based on the initial defect classifier, such as using the processor at 108.
- locations of interest on the next wafer are reviewed, such as using the image data acquisition system, based on the filtered inspection results and historical analysis sampling.
- the filtered inspection results are classified, such as using the processor, based on the reviewed locations of interest on the next wafer.
- the classified filtered results are stored in the central storage media.
- the defect classifier is updated, such as using the processor, based on the stored classified results in the central storage media.
- defects of interest in the next wafer are identified, such as using the processor, based on the classified filtered inspection results for the next wafer.
- Next wafer can refer to the next sequential wafer, but also can mean a second, third, fourth, fifth or later wafer.
- identifying defects of interest can include sampling near a classification boundary of a most recent defect classifier. Information can be obtained about classifier stability based on fluctuations in the defect classifier. Movement of the classification boundary can be predicted. Defects of interest can be identified based on the predicted movement in the classification boundary.
- the wafer inspection tool may perform a hot scan to capture inspection results using the method 100.
- the defect classifier may send defect of interest data and nuisance data to be used for retraining of the defect classifier.
- the defect classifier can be updated, such as using the processor, based on the stored classified results in the central storage media.
- the filtering step may be performed based on the updated defect classifier.
- Inspection results or reviewed locations of interest can be stored in the central storage media.
- Updating the defect classifier based on the stored classified results in the central storage media can include estimating a cap rate based on a calculated training confusion matrix.
- the calculated training confusion matrix may be based on the stored classified filtered inspection results for the next wafer in the central storage media.
- a nuisance rate can be estimated based on the defect classifier in the central storage media, the classified filtered inspection results for the next wafer, and the estimated cap rate.
- a confidence value also can be calculated based on the initial defect classifier.
- updating the defect classifier based on the stored classified results in the central storage media further can further include detecting a shadowing effect based on the defect classifier and the calculated confidence value.
- the filtered inspection results can have at least two thresholds associated with the filtered inspection results.
- a first of the at least two thresholds is for an inspection may be used for monitoring processes and defects.
- a second of the at least two thresholds is less than the first threshold and may be configured to capture sub-threshold defects during inspection. This enables sampling on both sides of the threshold to allow changing the classification boundary in both directions.
- This technique provides multiple advantages. It provides a fast cap rate estimator. Normally, the estimation of the cap rate is an expensive and/or inaccurate task. A user must sample a huge number of defects from a nuisance bin, review them with a tool (e.g., a SEM tool), classify them, and try to come up with an estimate of the number of DOI in the nuisance bin. This method is not feasible most of the time because the number of defects in the DOI bin is extremely large. Embodiments disclosed herein do not need any sample, which makes it extremely fast. A faster nuisance rate estimation is also provided. Normally to estimate the nuisance rate, the users randomly sample from the DOI bin and then SEM review them, and classify them. This extra time for sampling, SEM reviewing, and classification can be removed using techniques disclosed herein.
- a tool e.g., a SEM tool
- the estimate of the ROC curve on the entire wafer can be a helpful tool for the semiconductor manufacturers to tune the recipe and to identify the optimal conditions for the inspection given the desired outcome.
- the disclosed techniques also provide a detection method for shadowing effect.
- FIG. 9 is a block diagram of a system 200 for detecting defects of interest in a plurality of wafers.
- the system 200 includes a wafer inspection tool 201, an image data acquisition system 204, a central storage media 203, and a processor 202.
- the image data acquisition system 204 may be an SEM review tool.
- the wafer inspection tool 201 may be a BBP inspection tool, which can be configured to perform a hot scan to capture inspection results.
- the wafer inspection tool 201 also may be an LS tool or an unpatterned wafer surface inspection system, such as the Surfscan SPx manufactured by KLA-Tencor Corporation.
- the central storage media 203 is configured to store a plurality of classified inspection results and an initial defect classifier.
- the processor 202 is in electronic communication with the central storage media 203, the wafer inspection tool 201, and the image data acquisition system 204. [0067]
- the processor 202 is configured to execute the instructions of an inspection engine, a sampling engine, and a tuning engine.
- the inspection engine instructs the processor to receive inspection results of a first wafer from the wafer inspection tool.
- the sampling engine instructs the processor to: retrieve the initial defect classifier from the central storage media; filter the inspection results based on the initial defect classifier; review locations of interest on the first wafer from the image data acquisition system based on the filtered inspection results; classify the filtered inspection results based on the initial defect classifier; store the classified filtered inspection results in the central storage media; and identify defects of interest in the first wafer based on the classified filtered inspection results.
- the tuning engine instructs the processor to update the initial defect classifier based on the stored classified results in the central storage media. [0068] For each remaining wafer, the inspection engine instructs the processor to: receive inspection results of a next wafer from the wafer inspection tool.
- the sampling engine instructs the processor to: filter the inspection results based on the initial defect classifier; review locations of interest on the next wafer, using the image data acquisition system, based on the filtered inspection results and historical analysis sampling; classify the filtered inspection results based on the reviewed locations of interest on the next wafer; store the classified results in the central storage media; update the defect classifier, using the processor, based on the stored classified results in the central storage media; and identify defects of interest in the next wafer based on the classified filtered inspection results for the next wafer.
- the tuning engine can instruct the processor to update the defect classifier, using the processor, based on the stored classified results in the central storage media.
- the sampling engine can instruct the processor to perform the filtering step based on the updated defect classifier.
- the number of results or number of wafers used to update the defect classifier may be decided by the algorithm and can be controlled by setup. These numbers may depend on the use case and on the inspections. For research and development applications, only a few most recent wafers might be used. In a more mature high volume manufacturing process, the training data could come from more wafers. It may be time-bound and data- sufficiency bound.
- the defect classifier can send defect of interest data and nuisance data to be used for retraining of the defect classifier.
- the step of identifying defects of interest can include sampling near a
- classification boundary of a most recent defect classifier obtaining information about classifier stability based on fluctuations in the defect classifier; observing a movement in the classification boundary; and identifying defects of interest based on the predicted movement in the
- the inspection results or reviewed locations of interest can be stored in the central storage media 203, which may include a database.
- a central storage media 203 can store the classified defects along with the rest of the inspection population.
- a tuning and analysis engine can operate on the stored data after each new data is added to the database.
- a sampling engine can retrieve the latest classifier from the central server to identify the most suitable defects. This is done by one or more of the following techniques. First, leveraging the latest classifier to sample near the classification boundaries of the model (as both sides of the boundary). Second, using the information about classifier stability obtained from the classification fluctuations on recent wafers.
- processor 202 and central storage media 203 are illustrated as separate, these may be part of the same control unit. Both the processor 202 and central storage media 203 may be part of the wafer inspection tool 201 or the image data acquisition system 204, or another device. In an example, the processor 202 may be a standalone control unit or in a centralized quality control unit. Multiple processors 202 and/or central storage media 203 may be used. For example three processors 202 may be used for the inspection engine, sampling engine, and tuning engine. [0075] The processor 202 may be implemented in practice by any combination of hardware, software, and firmware.
- controller readable storage media such as a memory in the central storage media 203 or other memory.
- the processor 202 and central storage media 203 may be coupled to the components of the system 200 in any suitable manner (e.g., via one or more transmission media, which may include wired and/or wireless transmission media) such that the processor 202 and central storage media 203 can receive the output generated by the system 200.
- the processor 202 may be configured to perform a number of functions using the output.
- the processor 202 and central storage media 203, other system(s), or other subsystem(s) described herein may be part of various systems, including a personal computer system, image computer, mainframe computer system, workstation, network appliance, internet appliance, or other device.
- the subsystem(s) or system(s) may also include any suitable processor known in the art, such as a parallel processor.
- the subsystem(s) or system(s) may include a platform with high speed processing and software, either as a standalone or a networked tool.
- the different subsystems may be coupled to each other such that images, data, information, instructions, etc. can be sent between the subsystems.
- one subsystem may be coupled to additional subsystem(s) by any suitable transmission media, which may include any suitable wired and/or wireless transmission media known in the art.
- Two or more of such subsystems may also be effectively coupled by a shared computer-readable storage medium (not shown).
- An additional embodiment relates to a non-transitory computer-readable medium storing program instructions executable on a controller for performing a computer-implemented method of an embodiment disclosed herein.
- the processor 202 can be coupled to a memory in the central storage media 203 or other electronic data storage medium with non- transitory computer-readable medium that includes program instructions executable on the processor 202.
- the computer- implemented method may include any step(s) of any method(s) described herein.
- the processor 202 may be programmed to perform some or all of the steps of Figure 8.
- the memory in the central storage media 203 or other electronic data storage medium may be a storage medium such as a magnetic or optical disk, a magnetic tape, or any other suitable non-transitory computer-readable medium known in the art.
- the program instructions may be implemented in any of various ways, including procedure-based techniques, component-based techniques, and/or object-oriented techniques, among others.
- the program instructions may be implemented using ActiveX controls, C++ objects, JavaBeans, Microsoft Foundation Classes (MFC), SSE (Streaming SIMD Extension), or other technologies or methodologies, as desired.
- MFC Microsoft Foundation Classes
- SSE Streaming SIMD Extension
- each of the steps of the method may be performed as described herein.
- the methods also may include any other step(s) that can be performed by the controller and/or computer subsystem(s) or system(s) described herein.
- the steps can be performed by one or more computer systems, which may be configured according to any of the embodiments described herein.
- the methods described above may be performed by any of the system embodiments described herein.
Landscapes
- Engineering & Computer Science (AREA)
- Manufacturing & Machinery (AREA)
- Power Engineering (AREA)
- Microelectronics & Electronic Packaging (AREA)
- Computer Hardware Design (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Condensed Matter Physics & Semiconductors (AREA)
- Computing Systems (AREA)
- Data Mining & Analysis (AREA)
- Mathematical Optimization (AREA)
- Mathematical Analysis (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Evolutionary Computation (AREA)
- Pure & Applied Mathematics (AREA)
- Quality & Reliability (AREA)
- Automation & Control Theory (AREA)
- Computational Mathematics (AREA)
- Artificial Intelligence (AREA)
- Algebra (AREA)
- Probability & Statistics with Applications (AREA)
- Testing Or Measuring Of Semiconductors Or The Like (AREA)
- Investigating Materials By The Use Of Optical Means Adapted For Particular Applications (AREA)
Abstract
Wafer inspection with stable nuisance rates and defect of interest capture rates are disclosed. This technique can be used for discovery of newly appearing defects that occur during the manufacturing process. Based on a first wafer, defects of interest are identified based on the classified filtered inspection results. For each remaining wafer, the defect classifier is updated and defects of interest in the next wafer are identified based on the classified filtered inspection results.
Description
DIAGNOSTIC METHODS FOR THE CLASSIFIERS AND THE DEFECTS CAPTURED
BY OPTICAL TOOLS
Cross-Reference to Related Applications [0001] This application claims priority to U.S. Provisional Application No. 62/444,694 filed on January 10, 2017, U.S. Provisional Application No. 62/475,030 filed on March 22, 2017, and U.S. Provisional Application No. 62/581,378 filed November 3, 2017, the disclosures of which are incorporated herein by reference.
Field of the Disclosure [0002] The present disclosure relates to defect detection.
Background of the Disclosure
[0003] Evolution of the semiconductor manufacturing industry is placing ever greater demands on yield management and, in particular, on metrology and inspection systems. Critical dimensions are shrinking while wafer size is increasing. Economics is driving the industry to decrease the time for achieving high-yield, high-value production. Thus, minimizing the total time from detecting a yield problem to fixing it determines the return-on-investment for the semiconductor manufacturer.
[0004] Fabricating semiconductor devices, such as logic and memory devices, typically includes processing a semiconductor wafer using a large number of fabrication processes to form various features and multiple levels of the semiconductor devices. For example, lithography is a semiconductor fabrication process that involves transferring a pattern from a reticle to a photoresist arranged on a semiconductor wafer. Additional examples of semiconductor fabrication processes include, but are not limited to, chemical-mechanical polishing (CMP), etch, deposition, and ion implantation. Multiple semiconductor devices may be fabricated in an
arrangement on a single semiconductor wafer and then separated into individual semiconductor devices.
[0005] Algorithms can be used to detect defects on a wafer. When machine learning algorithms are used to create defect classifiers and nuisance filters, the algorithms tend to be considered as black box solutions that are not tuned or diagnosed. Assessment of an inspection recipe typically waited until observing a new set of labeled data received for the assessment or, alternatively, not using some part of the labeled data and reserving it for the validation. Both of these techniques wasted resources.
[0006] When setting up an inspection recipe, the overall performance assessment may be done based on the quality of the data used to train the classifier and the classifier's ability to learn and extract the information from the data. If the quality of the data is poor and the real defects and nuisance do not have a clear separation boundary, then any classifier will likely fail.
[0007] The performance of each recipe is evaluated using two measures: discriminability and reliability. There are many discriminability measures. One is the confusion matrix of the training data, which consists of a set of conditional error rates. From these conditional error rates, the cap rate and nuisance rate may be important to semiconductor manufacturers. Cap rate is the ratio of the number of defects of interest (DOI) that are classified correctly to the total number of DOI in the wafer. The nuisance rate is the number of nuisance that are classified as DOI to the total number of defects that are classified as DOI. A higher cap rate and a lower nuisance rate mean a better recipe. However, nuisance rate and cap rate could only previously be evaluated for the training data set that included the actual labels of data.
[0008] Reliability is a measure that shows how certain the classifier is about the decision it makes. It is a function of the estimation of posteriors done by the classifier. Previously, the classifier reliability was evaluated through the confidence calculation for each defect.
[0009] Although discriminability and reliability can be important measures, discriminability and reliability can cover up the reality if the underlying distributions of DOI and nuisance have certain characteristics. This can be referred to as a shadowing effect.
[0010] Commonly used approaches for the classifier assessment on broad band plasma (BBP) and laser scanning (LS) tools are based on the confusion matrix of the training set, which is used to measure discriminability, and calculating the confidence histogram, which is used to measure the reliability. As seen in Figure 1(a), the discriminability is only measured based on the training data, which is biased if any special sampling method being used. The confidence histogram has been used to measure reliability. There may be no information for a user about the reliability for the DOI or the nuisance classes using this technique.
[0011] Using the confusion matrix of the training set typically is not sufficient to understand the behavior of the recipe on the entire wafer. If the defects in the training set have been selected in certain way (which normally is done in order to reduce the number of defects for scanning electron microscope (SEM) review and manual classification), then the confusion matrix of the training set is biased toward those defects and is not be a good estimator of the classifier performance over the entire wafer.
[0012] Previous solutions retrain a binary classifier (e.g., nuisance vs. DOI) based on manual classification obtained during process monitoring (production sampling). These previous solutions used the updated classifier to create new DOI/nuisance separation on the subsequent wafer and used the new bins to generate production sample, which in turn was used to tune the next classifier. 50% of the previous solutions' sample is random sampling from the DOI bin of the latest classifier, and the other 50% is random sampling from the entire population. The two samples are used to compare the statistical process control (SPC) of the two inspections, and the second sample also provides "sub-threshold" defects to be used for retraining of the classifier. [0013] Another previous method for dealing with process/wafer variations relied on building the classifier from scratch and iteratively building the training set with the aid of the
SEM automatic defect classification (ADC), and then generating production samples from the new created DOI bin. However, the need to create a classifier from scratch on each wafer is more costly in terms of the SEM tool time. In addition, the ground truth for training the BBP model is based on SEM ADC without human verification, which makes the ground truth potentially less reliable. Lastly, this method does not leverage defects from previous wafers and, thus, increases the risk of data insufficiency and instability during the training process.
[0014] Previous techniques do not find the estimation of the cap rate and nuisance rate on the entire wafer (and for the unlabeled data) without additional sampling. Thus, the user does not know how tuning the recipe can affect the overall performance. Previous techniques also do not identify the shadowing effect. Therefore, a new defect detection technique and system is needed.
Brief Summary of the Disclosure
[0015] In a first embodiment, a system for detecting defects of interest in a plurality of wafers is provided. The system comprises a central storage media configured to store a plurality of classified inspection results and an initial defect classifier; a wafer inspection tool; an image data acquisition system; and a processor in electronic communication with the central storage media, the wafer inspection tool, and the image data acquisition system. The processor is configured to execute the instructions of: an inspection engine; a sampling engine, and a tuning engine. The inspection engine instructs the processor to receive inspection results of a first wafer from the wafer inspection tool. The sampling engine instructs the processor to: retrieve the initial defect classifier from the central storage media; filter the inspection results based on the initial defect classifier; review locations of interest on the first wafer from the image data acquisition system based on the filtered inspection results; classify the filtered inspection results based on the initial defect classifier; store the classified filtered inspection results in the central storage media; and identify defects of interest in the first wafer based on the classified filtered inspection results. The tuning engine instructs the processor to update the initial defect classifier based on the stored classified filtered inspection results in the central storage media. For each remaining wafer, the inspection engine instructs the processor to receive inspection results of a next wafer
from the wafer inspection tool. For each remaining wafer, the sampling engine instructs the processor to: filter the inspection results of the next wafer based on the initial defect classifier; review locations of interest on the next wafer, using the image data acquisition system based on the filtered inspection results of the next wafer and historical analysis sampling; classify the filtered inspection results of the next wafer based on the reviewed locations of interest on the next wafer; store the classified filtered inspection results for the next wafer in the central storage media; update the defect classifier, using the processor, based on the stored classified filtered inspection results for the next wafer in the central storage media; and identify defects of interest in the next wafer based on the classified filtered inspection results for the next wafer. [0016] For each of the remaining wafers, the tuning engine can instruct the processor to update the defect classifier, using the processor, based on the stored classified filtered inspection results for the next wafer in the central storage media. The sampling engine can instruct the processor to perform the filtering step based on the updated defect classifier.
[0017] The image data acquisition system can be an SEM review tool. [0018] The wafer inspection tool can perform a hot scan to capture inspection results.
For example, the wafer inspection tool can be a broadband plasma inspection tool.
[0019] The defect classifier can send defect of interest data and nuisance data for retraining of the defect classifier.
[0020] The step of identifying defects of interest can comprise: sampling near a classification boundary of a most recent defect classifier; obtaining information about classifier stability based on fluctuations in the defect classifier; observing a movement in the classification boundary; and identifying the defects of interest based on the predicted movement in the classification boundary.
[0021] The inspection results or reviewed locations of interest can be stored in the central storage media.
[0022] In a second embodiment, a method for identifying defects of interest in a plurality of wafers is provided. The method comprises receiving, at a processor, inspection results of a first wafer from a wafer inspection tool. Using the processor, the inspection results are filtered based on an initial defect classifier. Locations of interest on the first wafer are reviewed, using an image data acquisition system, based on the filtered inspection results. The filtered inspection results are classified, using the processor, based on the reviewed locations of interest on the first wafer. The classified filtered inspection results are stored in a central storage media. Defects of interest in the first wafer are identified based on the classified filtered inspection results. For each remaining wafer, the method comprises receiving, at the processor, inspection results of a next wafer from the wafer inspection tool. Using the processor, the inspection results are filtered based on the initial defect classifier. Locations of interest on the next wafer are reviewed, using the image data acquisition system, based on the filtered inspection results of the next wafer and historical analysis sampling. The filtered inspection results of the next wafer are classified, using the processor, based on the reviewed locations of interest on the next wafer. The classified filtered inspection results for the next wafer are stored in the central storage media. The defect classifier is updated, using the processor, based on the stored classified filtered inspection results for the next wafer in the central storage media. Defects of interest in the next wafer are identified based on the classified filtered inspection results for the next wafer.
[0023] The image data acquisition system can be an SEM review tool. [0024] The wafer inspection tool can perform a hot scan to capture inspection results. For example, the wafer inspection tool can be a broadband plasma inspection tool.
[0025] The defect classifier can send defect of interest data and nuisance data for retraining of the defect classifier.
[0026] The step of identifying defects of interest can comprise: sampling near a classification boundary of a most recent defect classifier; obtaining information about classifier stability based on fluctuations in the defect classifier; observing a movement in the classification
boundary; and identifying the defects of interest based on the predicted movement in the classification boundary.
[0027] For each of the remaining wafers, the method can comprise updating the defect classifier, using the processor, based on the stored classified filtered inspection results for the next wafer in the central storage media. The filtering step can be performed based on the updated defect classifier.
[0028] The inspection results or reviewed locations of interest can be stored in the central storage media.
[0029] The step of updating the defect classifier based on the stored classified filtered inspection results in the central storage media can comprise: estimating a cap rate based on a calculated training confusion matrix and estimating a nuisance rate based on the defect classifier in the central storage media, the classified filtered inspection results for the next wafer, and the estimated cap rate. The calculated training confusion matrix is based on the stored classified filtered inspection results for the next wafer in the central storage media. [0030] The filtered inspection results can have at least two thresholds associated with the filtered inspection results. A first of the at least two thresholds is for an inspection used for monitoring processes and defects. A second of the at least two thresholds is less than the first threshold and is configured to capture sub-threshold defects during inspection.
Description of the Drawings [0031] For a fuller understanding of the nature and objects of the disclosure, reference should be made to the following detailed description taken in conjunction with the
accompanying drawings, in which:
Figure 1 includes flowcharts (a) and (b) of previous techniques;
Figure 2 includes charts (a), (b), and (c) of a cap rate versus cutline curve, a nuisance rate versus cutline curve, and a cap rate versus nuisance rate curve, respectively;
Figure 3 includes distributions (a), (b), and (c);
Figure 4 includes distributions (a) and (b);
Figure 5 is a flowchart of an embodiment of a shadow detect algorithm in accordance with the present disclosure;
Figure 6 includes charts of accuracy versus number of defects in the pool for an ordinary wafer (a) and a shadowed wafer (b);
Figure 7 is a flowchart of an embodiment of a diagnostic model in accordance with the present disclosure;
Figure 8 is a flowchart of an embodiment in accordance with the present disclosure;
Figure 9 is a block diagram of a system in accordance with the present disclosure;
Figure 10 is a diagram of a dynamic classifier with dynamic sampling and stability analysis in accordance with the present disclosure; and
Figure 11 is a diagram of a static classifier with dynamic sampling and stability analysis in accordance with the present disclosure.
Detailed Description of the Disclosure
[0032] Although claimed subject matter will be described in terms of certain
embodiments, other embodiments, including embodiments that do not provide all of the benefits and features set forth herein, are also within the scope of this disclosure. Various structural, logical, process step, and electronic changes may be made without departing from the scope of the disclosure. Accordingly, the scope of the disclosure is defined only by reference to the appended claims.
[0033] The embodiments disclosed herein address new systems and methods for dealing with process and wafer instability in the early stages of an integrated circuit manufacturing process. One embodiment of the present disclosure is based on the idea of producing a small sample on productions lots in addition to the production sampling, aggregating the sample over
several wafers to build an up-to-date classifier, and using classifier to produce new updated samples on the next wafer.
[0034] The embodiments disclosed herein may be particularly advantageous over existing methods for at least the following reasons. The presently disclosed systems and methods utilize a supplemental (augmented) sample which is generated using the latest known process conditions and is well-suited for returning a superior classifier. The latest known process conditions and defects are far more useful for this purpose than the random samples currently used.
[0035] The latest known process conditions are also a superior indication of process change, and any new defects or defects that show the largest variation will show up efficiently in the sample. In other words, an efficient incremental discovery with a small sample size leads to smaller additional SEM review and classification costs.
[0036] Furthermore, with additional monitoring of the process instability, and the ability of the presently disclosed systems and methods to quantify that process instability, the supplemental sample can be automatically tuned to match those process conditions.
[0037] Generally, disclosed systems and methods allow for more relevant broadband plasma inspection with more stable nuisance rate and DOI capture rates. The disclosed systems and methods allow for faster discovery of newly appearing defects occurring during the manufacturing process, and allow for an analysis of the stability of the manufacturing process. [0038] There are several ways to implement the presently disclosed systems and methods. One embodiment relies solely on data from the central storage media and the systems and methods leverage the manual classification of defects in the classifier performance on the rest of the inspection. These embodiments cause the classifier to be one wafer behind what is currently being inspected. Another embodiment adds the ability to update the classifier on the current wafer by performing sampling on the wafer defect review tool and then producing the
supplemental sample the central storage media. One advantage of this embodiments is that the latest wafer condition is also included in the classifier.
[0039] The cap rate and the nuisance rate can be estimated for the data where the actual labels are not available. Thus, the expected values for the cap rate and the nuisance rate can be provided. The techniques show that all the estimations of cap rate, nuisance rate, posteriors, and confidence are accurate or that the data has the shadowed distribution. Data produced by algorithms (in addition to classification) can provide diagnostics information that is not available with manually created classifiers such as inline defect organizer (iDO). iDO is an example of an algorithm that can classify defects in real-time during inspection. [0040] A recipe can be assessed. These methods include estimation of the cap rate; estimation of the nuisance rate; evaluating the receiver operating curve (ROC), which can show cap rate versus nuisance rate, for fine tuning the recipe; and detecting the shadowing effect, which determines if the estimation of the posteriori, confidence, cap rate and nuisance rate are trustworthy. The ROC can be a curve that plots true positive rate versus false positive rate. Instead of or in collaboration with the ROC, DOI cap rate (true positive rate) versus nuisance rate (which is not false positive rate) may be used.
[0041] Two outputs from the classifier can be used to build the diagnostic tools. First, decisions, which are the classification results provided by the classifier, can be used. Second, posteriors for each defect can be used. There are different ways that a classifier might find posteriors. Distance from each class centroid or probabilistic measures of accuracy are two examples.
[0042] To estimate the cap rate, the ratio of the number of the DOIs that are correctly classified to the total number of DOIs in the training set can be used. This can be applied to the test set to find the estimation of the number of the DOI potentially missed in the test data.
Assuming that there are two classes (DOI and nuisance), the confusion matrix appears as shown in Table 1.
Table 1
Snm is the set of all defects that originally belongs to class m and classified as class n. SDD is the set of defects that are classified as DOI and are actually DOI. SND is the set of defects that are classified as nuisance and are actually DOI. SDN is the set of defects that are classified as DOI and are actually nuisance. SNN is the set of defects that are classified as nuisance and are also actually nuisance. The cap rate estimation for the entire wafer is shown in Equation 1.
In Equation 1 , |S| denotes the size (cardinality) of set S. To estimate the nuisance rate of unlabeled defects, a ratio of the accumulative posteriors of the defects associated to nuisance for the defects that are classified as DOI to the total number of defects in the DOI bin can be used. Assuming two classes (DOI and nuisance), the after classification of the population of the data for the test data (or any unlabeled data) is shown in Table 2.
Table 2
ADC
DOI SD
Nuisance SN
SD is the set of defects that are classified as DOI. SN is the set of defects that are classified as nuisance. Assuming the posterior for the nuisance class associated to defect i is pi, the nuisance rate would be calculated as shown in Equation 2.
Nu isance rate = — ;— ;—
i¾ i Eq. 2
[0043] SD is the set of defects that are classified as DOI. pi is the posterior probability for the nuisance class associated to defect i. |SD| denotes the size of set SD.
[0044] The cap rate can be increased with higher nuisance. This can be done by, for example, moving the cut lines in the confidence histogram and changing the class codes of the defects with lower confidence. The cap rate and nuisance rate can be evaluated for all the possible values of the cutline. Then the three plots can be shown, three examples of which are shown in Figure 2. Chart (a) in Figure 2 shows cap rate versus cutline value. Chart (b) in Figure 2 shows nuisance rate versus cutline value. Chart (c) in Figure 2 shows ROC. An ROC can be a useful representation of the performance of a classifier on a given data set. The user can find what is going to be the nuisance rate for a desired capture rate, and vice versa. With these curves, a user can decide if the values of cutline worthwhile.
[0045] In classification, well-separated distributions may be the ones with short overlap as shown in Figure 3(a), (b), and (c). Data can be well-separated as if a clear boundary can be draw between the two distributions as shown in Figure 3 (a) and (b). Distributions can be well- separated and have multiple regions in the space and be separated using multiple boundaries as shown in Figure 3(c).
[0046] Most classifiers can learn this situation. In this scenario, the performance of the classifier is ordinary. Such Probability Density Functions (PDFs) are the ones that normally appears in the wafer, but this is not always the case. A big part of one distribution may have been shadowed. Shadowing effect is a situation when a large part of one class distribution is under the PDF of another class. This situation can happen as mistakes during manual or automated labeling or as a result of not having good attributes to distinguish the shadowed part from the other class. Charts (a) and (b) in Figure 4 are two examples of this situation.
[0047] The detection of the first case ((a) in Figure 4) is relatively easy because, just by observing the training confusion matrix, it can be determined that the accuracy for one class is poor. Detecting the second case ((b) in Figure 4) is more difficult. This situation can mislead a user about the data on the wafer where big portion of one class will not be detected, no matter what kind of the classifier is used. The misclassification here is not due to the poor performance of the classifier, but may be due to the poor quality of features or labeling.
[0048] To detect this situation, a classifier can be trained with the training set. Then, the training can be sorted to set ascendingly from the confidence values the defects obtained from the classifier. An empty pool can be created and defects can be added one by one from lowest confidence to highest confidence to the pool. After adding each defect, the confusion matrix of the defects in the pool can be calculated and the accuracies of the classes and the number of defects in the pool can be saved. Accuracy for each class can be defined as the number of correctly classified defects of that class to the total number of defects from that class. After using all the defects in the training set, the accuracies versus the number of defects in the pool can be compared. An example of this algorithm is shown in Figure 5.
[0049] For an ordinary defect distribution on a wafer, the accuracy of all of the classes in the pool is expected to increase or stay constant when the number of defects is increased. While other rationales are possible, a new defect in the pool may have greater than or equal confidence in comparison to the previous defects in the pool.
[0050] The plots in (a) and (b) in Figure 6 show this for two different wafers. The plot in
(a) of Figure 6 is from a wafer without shadowed DOI and both DOI and nuisance accuracies improve with the number of defects. However, the plot (b) in Figure 6 shows a wafer which a DOI class observes a shadowing effect. The DOI bin does not improve with the number of defects. It indicates that high confidence defect are added, but these are being classified incorrectly, which is an indication of the shadowing effect.
[0051] The details of an embodiment of a method are shown in the flowchart of Figure 7.
Figure 7 shows a flow chart for the algorithm of estimating nuisance rate and capture rate and detection of the shadow effect. The training set is used to create the classifier. The classifier is applied to the defects in the test set. Then, the classifier is used to evaluate the confidence and posteriors for all the defects (both in the training set and test set). The estimation of the nuisance rate is done using the posteriors. The estimation of the capture rate is done using the confusion matrix obtained from the training set. Finally, a check is done to find out if the data is under the shadow effect or not. If it is not, then the estimates are trustable. [0052] Figure 8 is a flowchart of a method 100 for identifying defects of interest in a plurality of wafers. At 101, inspection results of a first wafer are received, such as at a processor, from a wafer inspection tool, which may be a BBP tool or another inspection device. At 102, the inspection results are filtered based on an initial defect classifier, such as using the processor. At 103, locations of interest on the first wafer are reviewed based on the filtered inspection results, such as using an image data acquisition system. The image data acquisition system may be an SEM review tool or another measurement, inspection, or metrology tool. At 104, the filtered inspection results are classified, such as using the processor, based on the reviewed locations of interest on the first wafer. At 105, the classified inspection results are stored in a central storage media. At 106, defects of interest are identified based on the classified filtered inspection results, such as using the processor. Filtered inspection results may be kept separate, such as for each wafer that is sampled.
[0053] For each remaining wafer, inspection results of the next wafer are received, such as at the processes, from the wafer inspection tool at 107. The inspection results are filtered based on the initial defect classifier, such as using the processor at 108. At 109, locations of interest on the next wafer are reviewed, such as using the image data acquisition system, based on the filtered inspection results and historical analysis sampling. At 110, the filtered inspection results are classified, such as using the processor, based on the reviewed locations of interest on the next wafer. At 111, the classified filtered results are stored in the central storage media. At 112, the defect classifier is updated, such as using the processor, based on the stored classified results in the central storage media. At 113, defects of interest in the next wafer are identified, such as using the processor, based on the classified filtered inspection results for the next wafer.
[0054] Next wafer can refer to the next sequential wafer, but also can mean a second, third, fourth, fifth or later wafer.
[0055] In the method 100, identifying defects of interest can include sampling near a classification boundary of a most recent defect classifier. Information can be obtained about classifier stability based on fluctuations in the defect classifier. Movement of the classification boundary can be predicted. Defects of interest can be identified based on the predicted movement in the classification boundary.
[0056] The wafer inspection tool may perform a hot scan to capture inspection results using the method 100. [0057] The defect classifier may send defect of interest data and nuisance data to be used for retraining of the defect classifier.
[0058] For each remaining wafer, the defect classifier can be updated, such as using the processor, based on the stored classified results in the central storage media. The filtering step may be performed based on the updated defect classifier.
[0059] Inspection results or reviewed locations of interest can be stored in the central storage media.
[0060] Updating the defect classifier based on the stored classified results in the central storage media can include estimating a cap rate based on a calculated training confusion matrix. The calculated training confusion matrix may be based on the stored classified filtered inspection results for the next wafer in the central storage media. A nuisance rate can be estimated based on the defect classifier in the central storage media, the classified filtered inspection results for the next wafer, and the estimated cap rate. These steps may be performed by the processor.
[0061] A confidence value also can be calculated based on the initial defect classifier. In this instance, updating the defect classifier based on the stored classified results in the central storage media further can further include detecting a shadowing effect based on the defect classifier and the calculated confidence value.
[0062] The filtered inspection results can have at least two thresholds associated with the filtered inspection results. A first of the at least two thresholds is for an inspection may be used for monitoring processes and defects. A second of the at least two thresholds is less than the first threshold and may be configured to capture sub-threshold defects during inspection. This enables sampling on both sides of the threshold to allow changing the classification boundary in both directions.
[0063] This technique provides multiple advantages. It provides a fast cap rate estimator. Normally, the estimation of the cap rate is an expensive and/or inaccurate task. A user must sample a huge number of defects from a nuisance bin, review them with a tool (e.g., a SEM tool), classify them, and try to come up with an estimate of the number of DOI in the nuisance bin. This method is not feasible most of the time because the number of defects in the DOI bin is extremely large. Embodiments disclosed herein do not need any sample, which makes it extremely fast. A faster nuisance rate estimation is also provided. Normally to estimate the nuisance rate, the users randomly sample from the DOI bin and then SEM review them, and
classify them. This extra time for sampling, SEM reviewing, and classification can be removed using techniques disclosed herein.
[0064] The estimate of the ROC curve on the entire wafer can be a helpful tool for the semiconductor manufacturers to tune the recipe and to identify the optimal conditions for the inspection given the desired outcome.
[0065] The disclosed techniques also provide a detection method for shadowing effect.
Inseparable parts of the distribution in the data can be identified. This phenomenon normally happens due to the mistakes during manual labeling, poor SEM image quality, or lack of strong features. [0066] Figure 9 is a block diagram of a system 200 for detecting defects of interest in a plurality of wafers. The system 200 includes a wafer inspection tool 201, an image data acquisition system 204, a central storage media 203, and a processor 202. The image data acquisition system 204 may be an SEM review tool. The wafer inspection tool 201 may be a BBP inspection tool, which can be configured to perform a hot scan to capture inspection results. The wafer inspection tool 201 also may be an LS tool or an unpatterned wafer surface inspection system, such as the Surfscan SPx manufactured by KLA-Tencor Corporation. The central storage media 203 is configured to store a plurality of classified inspection results and an initial defect classifier. The processor 202 is in electronic communication with the central storage media 203, the wafer inspection tool 201, and the image data acquisition system 204. [0067] The processor 202 is configured to execute the instructions of an inspection engine, a sampling engine, and a tuning engine. The inspection engine instructs the processor to receive inspection results of a first wafer from the wafer inspection tool. The sampling engine instructs the processor to: retrieve the initial defect classifier from the central storage media; filter the inspection results based on the initial defect classifier; review locations of interest on the first wafer from the image data acquisition system based on the filtered inspection results; classify the filtered inspection results based on the initial defect classifier; store the classified
filtered inspection results in the central storage media; and identify defects of interest in the first wafer based on the classified filtered inspection results. The tuning engine instructs the processor to update the initial defect classifier based on the stored classified results in the central storage media. [0068] For each remaining wafer, the inspection engine instructs the processor to: receive inspection results of a next wafer from the wafer inspection tool. The sampling engine instructs the processor to: filter the inspection results based on the initial defect classifier; review locations of interest on the next wafer, using the image data acquisition system, based on the filtered inspection results and historical analysis sampling; classify the filtered inspection results based on the reviewed locations of interest on the next wafer; store the classified results in the central storage media; update the defect classifier, using the processor, based on the stored classified results in the central storage media; and identify defects of interest in the next wafer based on the classified filtered inspection results for the next wafer.
[0069] For each remaining wafer, the tuning engine can instruct the processor to update the defect classifier, using the processor, based on the stored classified results in the central storage media. The sampling engine can instruct the processor to perform the filtering step based on the updated defect classifier. The number of results or number of wafers used to update the defect classifier may be decided by the algorithm and can be controlled by setup. These numbers may depend on the use case and on the inspections. For research and development applications, only a few most recent wafers might be used. In a more mature high volume manufacturing process, the training data could come from more wafers. It may be time-bound and data- sufficiency bound.
[0070] The defect classifier can send defect of interest data and nuisance data to be used for retraining of the defect classifier. [0071] The step of identifying defects of interest can include sampling near a
classification boundary of a most recent defect classifier; obtaining information about classifier
stability based on fluctuations in the defect classifier; observing a movement in the classification boundary; and identifying defects of interest based on the predicted movement in the
classification boundary. Observing a movement may be performed on some of the most recent wafers. [0072] The inspection results or reviewed locations of interest can be stored in the central storage media 203, which may include a database. In a particular instance, a central storage media 203 can store the classified defects along with the rest of the inspection population. A tuning and analysis engine can operate on the stored data after each new data is added to the database. A sampling engine can retrieve the latest classifier from the central server to identify the most suitable defects. This is done by one or more of the following techniques. First, leveraging the latest classifier to sample near the classification boundaries of the model (as both sides of the boundary). Second, using the information about classifier stability obtained from the classification fluctuations on recent wafers. Third, directing most of the sample to the side of the classification boundary that is most likely be in the direction of the boundary move. [0073] One embodiment relies solely on the central storage media 203 and leverages the manual classification of defects and the classifier performance on the rest of the inspection. This construction keeps the classifier one wafer behind. Another embodiment adds the ability to update the model on the current wafer by performing sampling on the wafer defect review tool and then producing the supplemental sample for central storage, which means that the latest wafer condition is also included. Two examples are shown in Figures 10 and 11. In Figures 10 and 11, inspection is run hotter with a nuisance-DOI classifier playing the role of a standard nuisance filter. This retains defects on both sides of the nuisance-DOI boundary for retraining. Stability information from historical analysis sampling setup and the latest classifier is used for sampling. [0074] While the processor 202 and central storage media 203 are illustrated as separate, these may be part of the same control unit. Both the processor 202 and central storage media 203 may be part of the wafer inspection tool 201 or the image data acquisition system 204, or another
device. In an example, the processor 202 may be a standalone control unit or in a centralized quality control unit. Multiple processors 202 and/or central storage media 203 may be used. For example three processors 202 may be used for the inspection engine, sampling engine, and tuning engine. [0075] The processor 202 may be implemented in practice by any combination of hardware, software, and firmware. Also, its functions as described herein may be performed by one unit, or divided up among different components, each of which may be implemented in turn by any combination of hardware, software and firmware. Program code or instructions for the processor 202 to implement various methods and functions may be stored in controller readable storage media, such as a memory in the central storage media 203 or other memory.
[0076] The processor 202 and central storage media 203 may be coupled to the components of the system 200 in any suitable manner (e.g., via one or more transmission media, which may include wired and/or wireless transmission media) such that the processor 202 and central storage media 203 can receive the output generated by the system 200. The processor 202 may be configured to perform a number of functions using the output.
[0077] The processor 202 and central storage media 203, other system(s), or other subsystem(s) described herein may be part of various systems, including a personal computer system, image computer, mainframe computer system, workstation, network appliance, internet appliance, or other device. The subsystem(s) or system(s) may also include any suitable processor known in the art, such as a parallel processor. In addition, the subsystem(s) or system(s) may include a platform with high speed processing and software, either as a standalone or a networked tool.
[0078] If the system includes more than one subsystem, then the different subsystems may be coupled to each other such that images, data, information, instructions, etc. can be sent between the subsystems. For example, one subsystem may be coupled to additional subsystem(s) by any suitable transmission media, which may include any suitable wired and/or wireless
transmission media known in the art. Two or more of such subsystems may also be effectively coupled by a shared computer-readable storage medium (not shown).
[0079] An additional embodiment relates to a non-transitory computer-readable medium storing program instructions executable on a controller for performing a computer-implemented method of an embodiment disclosed herein. In particular, the processor 202 can be coupled to a memory in the central storage media 203 or other electronic data storage medium with non- transitory computer-readable medium that includes program instructions executable on the processor 202. The computer- implemented method may include any step(s) of any method(s) described herein. For example, the processor 202 may be programmed to perform some or all of the steps of Figure 8. The memory in the central storage media 203 or other electronic data storage medium may be a storage medium such as a magnetic or optical disk, a magnetic tape, or any other suitable non-transitory computer-readable medium known in the art.
[0080] The program instructions may be implemented in any of various ways, including procedure-based techniques, component-based techniques, and/or object-oriented techniques, among others. For example, the program instructions may be implemented using ActiveX controls, C++ objects, JavaBeans, Microsoft Foundation Classes (MFC), SSE (Streaming SIMD Extension), or other technologies or methodologies, as desired.
[0081] Each of the steps of the method may be performed as described herein. The methods also may include any other step(s) that can be performed by the controller and/or computer subsystem(s) or system(s) described herein. The steps can be performed by one or more computer systems, which may be configured according to any of the embodiments described herein. In addition, the methods described above may be performed by any of the system embodiments described herein.
[0082] Although the present disclosure has been described with respect to one or more particular embodiments, it will be understood that other embodiments of the present disclosure may be made without departing from the spirit and scope of the present disclosure. Hence, the
present disclosure is deemed limited only by the appended claims and the reasonable interpretation thereof.
Claims
1. A system for detecting defects of interest in a plurality of wafers comprising:
a central storage media configured to store a plurality of classified inspection results and an initial defect classifier;
a wafer inspection tool;
an image data acquisition system; and
a processor in electronic communication with the central storage media, the wafer inspection tool, and the image data acquisition system, the processor configured to execute the instructions of:
an inspection engine which instructs the processor to:
receive inspection results of a first wafer from the wafer inspection tool; a sampling engine which instructs the processor to:
retrieve the initial defect classifier from the central storage media;
filter the inspection results based on the initial defect classifier;
review locations of interest on the first wafer from the image data acquisition system, based on the filtered inspection results;
classify the filtered inspection results based on the initial defect classifier; store the classified filtered inspection results in the central storage media; and identify defects of interest in the first wafer based on the classified filtered inspection results;
a tuning engine which instructs the processor to:
update the initial defect classifier based on the stored classified filtered inspection results in the central storage media;
wherein for each remaining wafer:
the inspection engine instructs the processor to:
receive inspection results of a next wafer from the wafer inspection tool; the sampling engine instructs the processor to:
filter the inspection results of the next wafer based on the initial defect classifier;
review locations of interest on the next wafer, using the image data acquisition system based on the filtered inspection results of the next wafer and historical analysis sampling;
classify the filtered inspection results of the next wafer based on the reviewed
locations of interest on the next wafer;
store the classified filtered inspection results for the next wafer in the central storage media;
update the defect classifier, using the processor, based on the stored classified filtered inspection results for the next wafer in the central storage media; and identify defects of interest in the next wafer based on the classified filtered inspection results for the next wafer.
2. The system of claim 1, wherein, for each of the remaining wafers:
the tuning engine instructs the processor to:
update the defect classifier, using the processor, based on the stored classified filtered inspection results for the next wafer in the central storage media;
wherein the sampling engine instructs the processor to perform the filtering step based on the updated defect classifier.
3. The system of claim 1, wherein the image data acquisition system is an SEM review tool.
4. The system of claim 1, wherein the wafer inspection tool performs a hot scan to capture inspection results.
5. The system of claim 1, wherein the defect classifier sends defect of interest data and nuisance data for retraining of the defect classifier.
6. The system of claim 1, wherein the step of identifying defects of interest comprises:
sampling near a classification boundary of a most recent defect classifier;
obtaining information about classifier stability based on fluctuations in the defect classifier;
observing a movement in the classification boundary; and
identifying the defects of interest based on the predicted movement in the classification boundary.
7. The system of claim 1, further comprising storing the inspection results or reviewed locations of interest in the central storage media.
8. The system of claim 1, wherein the wafer inspection tool is a broadband plasma inspection tool.
9. A method for identifying defects of interest in a plurality of wafers comprising:
receiving, at a processor, inspection results of a first wafer from a wafer inspection tool; filtering, using the processor, the inspection results based on an initial defect classifier; reviewing locations of interest on the first wafer, using an image data acquisition system, based on the filtered inspection results;
classifying the filtered inspection results, using the processor, based on the reviewed
locations of interest on the first wafer;
storing the classified filtered inspection results in a central storage media;
identifying defects of interest in the first wafer based on the classified filtered inspection results; and
for each remaining wafer:
receiving, at the processor, inspection results of a next wafer from the wafer inspection tool;
filtering, using the processor, the inspection results based on the initial defect classifier; reviewing locations of interest on the next wafer, using the image data acquisition
system, based on the filtered inspection results of the next wafer and historical analysis sampling;
classifying the filtered inspection results of the next wafer, using the processor, based on the reviewed locations of interest on the next wafer;
storing the classified filtered inspection results for the next wafer in the central storage media;
updating the defect classifier, using the processor, based on the stored classified filtered inspection results for the next wafer in the central storage media; and
identifying defects of interest in the next wafer based on the classified filtered inspection results for the next wafer.
10. The method of claim 9, wherein the image data acquisition system is an SEM review tool.
11. The method of claim 9, wherein the wafer inspection tool performs a hot scan to capture inspection results.
12. The method of claim 9, wherein the defect classifier sends defect of interest data and nuisance data for retraining of the defect classifier.
13. The method of claim 9, wherein the step of identifying defects of interest comprises:
sampling near a classification boundary of a most recent defect classifier;
obtaining information about classifier stability based on fluctuations in the defect classifier; observing a movement in the classification boundary; and
identifying the defects of interest based on the predicted movement in the classification boundary.
14. The method of claim 9, further comprising, for each of the remaining wafers:
updating the defect classifier, using the processor, based on the stored classified filtered inspection results for the next wafer in the central storage media;
wherein the filtering step is performed based on the updated defect classifier.
15. The method of claim 9, further comprising storing the inspection results or reviewed locations of interest in the central storage media.
16. The method of claim 9, wherein the wafer inspection tool is a broadband plasma inspection tool.
17. The method of claim 9, wherein the step of updating the defect classifier based on the stored classified filtered inspection results in the central storage media comprises:
estimating a cap rate based on a calculated training confusion matrix, wherein the calculated training confusion matrix is based on the stored classified filtered inspection results for the next wafer in the central storage media; and
estimating a nuisance rate based on the defect classifier in the central storage media, the classified filtered inspection results for the next wafer, and the estimated cap rate.
18. The method of claim 9, wherein the filtered inspection results have at least two thresholds associated with the filtered inspection results, wherein a first of the at least two thresholds is for an inspection used for monitoring processes and defects, and wherein a second of the at least two thresholds is less than the first threshold and is configured to capture sub-threshold defects during inspection.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020197023546A KR102557181B1 (en) | 2017-01-10 | 2018-01-05 | Diagnostic methods for faults captured by optical tools and classifiers |
Applications Claiming Priority (10)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201762444694P | 2017-01-10 | 2017-01-10 | |
US62/444,694 | 2017-01-10 | ||
IN201741004615 | 2017-02-08 | ||
IN201741004615 | 2017-02-08 | ||
US201762475030P | 2017-03-22 | 2017-03-22 | |
US62/475,030 | 2017-03-22 | ||
US201762581378P | 2017-11-03 | 2017-11-03 | |
US62/581,378 | 2017-11-03 | ||
US15/835,399 US11237119B2 (en) | 2017-01-10 | 2017-12-07 | Diagnostic methods for the classifiers and the defects captured by optical tools |
US15/835,399 | 2017-12-07 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2018132321A1 true WO2018132321A1 (en) | 2018-07-19 |
Family
ID=62840218
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2018/012684 WO2018132321A1 (en) | 2017-01-10 | 2018-01-05 | Diagnostic methods for the classifiers and the defects captured by optical tools |
Country Status (3)
Country | Link |
---|---|
KR (1) | KR102557181B1 (en) |
TW (1) | TWI755468B (en) |
WO (1) | WO2018132321A1 (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113169089A (en) * | 2018-12-18 | 2021-07-23 | 科磊股份有限公司 | Optical Mode Selection for Multimode Semiconductor Inspection |
CN114174812A (en) * | 2019-08-01 | 2022-03-11 | 科磊股份有限公司 | Method for process monitoring with optical inspection |
CN115088125A (en) * | 2020-03-20 | 2022-09-20 | 舍弗勒技术股份两合公司 | Method and inspection device for inspecting bipolar plates of electrochemical cells, in particular fuel cells |
CN115951619A (en) * | 2023-03-09 | 2023-04-11 | 山东拓新电气有限公司 | Remote intelligent control system of heading machine based on artificial intelligence |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6256093B1 (en) * | 1998-06-25 | 2001-07-03 | Applied Materials, Inc. | On-the-fly automatic defect classification for substrates using signal attributes |
US20050177264A1 (en) * | 2004-02-06 | 2005-08-11 | Long-Hui Lin | Method of building a defect database |
US20120229618A1 (en) * | 2009-09-28 | 2012-09-13 | Takahiro Urano | Defect inspection device and defect inspection method |
US20140050389A1 (en) * | 2012-08-14 | 2014-02-20 | Kla-Tencor Corporation | Automated Inspection Scenario Generation |
US20160258879A1 (en) * | 2015-03-03 | 2016-09-08 | Kla-Tencor Corporation | Adaptive Nuisance Filter |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8126255B2 (en) | 2007-09-20 | 2012-02-28 | Kla-Tencor Corp. | Systems and methods for creating persistent data for a wafer and for using persistent data for inspection-related functions |
US9430743B2 (en) * | 2014-03-06 | 2016-08-30 | Kla-Tencor Corp. | Composite defect classifier |
-
2018
- 2018-01-05 WO PCT/US2018/012684 patent/WO2018132321A1/en active Application Filing
- 2018-01-05 KR KR1020197023546A patent/KR102557181B1/en active Active
- 2018-01-09 TW TW107100714A patent/TWI755468B/en active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6256093B1 (en) * | 1998-06-25 | 2001-07-03 | Applied Materials, Inc. | On-the-fly automatic defect classification for substrates using signal attributes |
US20050177264A1 (en) * | 2004-02-06 | 2005-08-11 | Long-Hui Lin | Method of building a defect database |
US20120229618A1 (en) * | 2009-09-28 | 2012-09-13 | Takahiro Urano | Defect inspection device and defect inspection method |
US20140050389A1 (en) * | 2012-08-14 | 2014-02-20 | Kla-Tencor Corporation | Automated Inspection Scenario Generation |
US20160258879A1 (en) * | 2015-03-03 | 2016-09-08 | Kla-Tencor Corporation | Adaptive Nuisance Filter |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113169089A (en) * | 2018-12-18 | 2021-07-23 | 科磊股份有限公司 | Optical Mode Selection for Multimode Semiconductor Inspection |
CN113169089B (en) * | 2018-12-18 | 2024-04-09 | 科磊股份有限公司 | Optical Mode Selection for Multimode Semiconductor Inspection |
CN114174812A (en) * | 2019-08-01 | 2022-03-11 | 科磊股份有限公司 | Method for process monitoring with optical inspection |
CN114174812B (en) * | 2019-08-01 | 2022-12-16 | 科磊股份有限公司 | Method for process monitoring with optical inspection |
CN115088125A (en) * | 2020-03-20 | 2022-09-20 | 舍弗勒技术股份两合公司 | Method and inspection device for inspecting bipolar plates of electrochemical cells, in particular fuel cells |
CN115951619A (en) * | 2023-03-09 | 2023-04-11 | 山东拓新电气有限公司 | Remote intelligent control system of heading machine based on artificial intelligence |
Also Published As
Publication number | Publication date |
---|---|
KR20190098271A (en) | 2019-08-21 |
TWI755468B (en) | 2022-02-21 |
KR102557181B1 (en) | 2023-07-18 |
TW201830334A (en) | 2018-08-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11237119B2 (en) | Diagnostic methods for the classifiers and the defects captured by optical tools | |
US10789703B2 (en) | Semi-supervised anomaly detection in scanning electron microscope images | |
US12007335B2 (en) | Automatic optimization of an examination recipe | |
CN118196555B (en) | Generating training data usable for inspection of semiconductor samples | |
TWI571946B (en) | Iterative defect filtering process | |
US10267748B2 (en) | Optimizing training sets used for setting up inspection-related algorithms | |
CN109598698B (en) | System, method, and non-transitory computer readable medium for classifying a plurality of items | |
JP2018128468A (en) | Optimization of unknown defect removal for automatic defect classification | |
TW201915770A (en) | Method of classifying defects in a semiconductor specimen and system thereof | |
WO2018132321A1 (en) | Diagnostic methods for the classifiers and the defects captured by optical tools | |
JP2013224943A (en) | Integration of automatic and manual defect classification | |
CN112805719B (en) | Classifying defects in semiconductor samples | |
EP4107593B1 (en) | Determination of defect location for examination of a specimen | |
KR102360769B1 (en) | Automatic selection of algorithmic modules for examination of a specimen | |
KR102685359B1 (en) | Method of examining specimens and system thereof | |
US12141173B2 (en) | Error factor estimation device and error factor estimation method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 18738721 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 20197023546 Country of ref document: KR Kind code of ref document: A |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 18738721 Country of ref document: EP Kind code of ref document: A1 |