Computational lithography

From Wikipedia, the free encyclopedia
Jump to: navigation, search

Computational lithography (also known as computational scaling) is the set of mathematical and algorithmic approaches designed to improve the resolution attainable through photolithography. Computational lithography has come to the forefront of photolithography in 2008 as the semiconductor industry grappled with the challenges associated with the transition to 22 nanometer CMOS process technology and beyond.

Context: industry forced to extend 193nm deep UV photolithography[edit]

The periodic enhancement in the resolution achieved through photolithography has been a driving force behind Moore's Law. Resolution improvements enable printing of smaller geometries on an integrated circuit. The minimum feature size that a projection system typically used in photolithography can print is given approximately by:

CD = k_1 \cdot\frac{\lambda}{NA}

where

\,CD is the minimum feature size (also called the critical dimension).

\,\lambda is the wavelength of light used.

\,NA is the numerical aperture of the lens as seen from the wafer.

\,k_1 (commonly called k1 factor) is a coefficient that encapsulates process-related factors.

Historically, resolution enhancements in photolithography have been achieved through the progression of stepper illumination sources to smaller and smaller wavelengths — from "g-line" (436 nm) and "i-line" (365 nm) sources based on mercury lamps, to the current systems based on deep ultraviolet excimer lasers sources at 193 nm. However the progression to yet finer wavelength sources has been stalled by the intractable problems associated with extreme ultraviolet lithography and x-ray lithography, forcing semiconductor manufacturers to extend the current 193 nm optical lithography systems until some form of next-generation lithography proves viable (although 157 nm steppers have also been marketed, they have proven cost-prohibitive at $50M each).[1] Efforts to improve resolution by increasing the numerical aperture have led to the use of immersion lithography. As further improvements in resolution through wavelength reduction or increases in numerical aperture have become either technically challenging or economically unfeasible, much attention has been paid to reducing the k1-factor. The k1 factor can be reduced through process improvements, such as phase-shift photomasks. These techniques have enabled photolithography at the 32 nanometer CMOS process technology node using a wavelength of 193 nm (deep ultraviolet). However, with the ITRS roadmap calling for the 22 nanometer node to be in use by 2011, photolithography researchers have had to develop an additional suite of improvements to make 22 nm technology manufacturable.[1] While the increase in mathematical modeling has been underway for some time, the degree and expense of those calculations has justified the use of a new term to cover the changing landscape: computational lithography.

A short history of computational lithography[edit]

Computational Lithography means the use of computers to simulate printing of micro-lithography structures. Pioneering work was done by Chris Mack at NSA in developing PROLITH, Rick Dill at IBM and Andy Neureuther at University of California, Berkeley from the early 1980s. These tools were limited to lithography process optimization as the algorithms were limited to a few square micrometres of resist. Commercial full-chip optical proximity correction, using model forms, was first implemented by TMA (now a subsidiary of Synopsys) and Numerical Technologies (also part of Synopsys) around 1997.[2] Since then the market and complexity has grown significantly. With the move to sub-wavelength lithography at the 180 nm and 130 nm nodes, RET techniques such as Assist features, Phase Shift Masks started to be used together with OPC. For the transition from 65 nm to 45 nm nodes customers were worrying that not only that design rules were insufficient to guarantee printing without yield limiting hotspots, but also that tape-out time may need thousands of CPUs or weeks of run time. This predicted exponential increase in computational complexity for mask synthesis on moving to the 45 nm process node spawned a significant venture capital investment in Design for Manufacturing start-up companies.[3] A number of startup companies promoting their own disruptive solutions to this problem started to appear, techniques from custom hardware acceleration to radical new algorithms such as Inverse Lithography were touted to resolve the forthcoming bottlenecks. Despite all this activity, incumbent OPC suppliers were able to adapt and keep their major customers, with RET and OPC being used together as for previous nodes, but now on more layers and with larger data files, and turn around time concerns were met by new algorithms and improvements in multi-core commodity processors. The term computational lithography was first used by Brion Technology (now a subsidiary of ASML) in 2005[4] to promote their hardware accelerated full chip lithography simulation platform. Since then the term has been used by the industry to describe full chip mask synthesis solutions. As 45 nm goes into full production and EUV lithography introduction is delayed, 32 nm and 22 nm are expected to run on existing 193 nm scanners technology. Now, not only are throughput and capabilities concerns resurfacing, but also new computational lithography techniques such as Source Mask Optimization (SMO) is seen as a way to squeeze better resolution specific to a given design. Today, all the major Mask Synthesis vendors have settled on the term "Computational Lithography" to describe and promote the set of Mask Synthesis technologies required for 22 nm.

Techniques comprising computational lithography[edit]

Computational lithography makes use of a number of numerical simulations to improve the performance (resolution and contrast) of cutting-edge photomasks. The combined techniques include Resolution Enhancement Technology (RET), Optical Proximity Correction (OPC), Source Mask Optimization (SMO), etc.[5] The techniques vary in terms of their technical feasibility and engineering sensible-ness, resulting in the adoption of some and the continual R&D of others.[6]

Resolution Enhancement Technology (RET)[edit]

Resolution Enhancement Technology, first used in the 90 nanometer generation, using the mathematics of diffraction optics to specify multi-layer phase-shift photomasks that use interference patterns in the photomask that enhance resolution on the printed wafer surface.

Optical Proximity Correction (OPC)[edit]

Optical proximity correction uses computational methods to counteract the effects of diffraction-related blurring and under-exposure by modifying on-mask geometries with means such as:

  • adjusting linewidths depending on the density of surrounding geometries (a trace surrounded by a large open area will be over-exposed compared with the same trace surrounded by a dense pattern)
  • adding "dog-bone" endcaps to the end of lines to prevent line shortening
  • correcting for electron beam proximity effects

OPC can be broadly divided into rule-based and model-based.[7] Inverse lithography technology, which treats the OPC as an inverse imaging problem, is also a useful technique because it can provide unintuitive mask patterns.[8]

Complex modeling of the lens system and photoresist[edit]

Beyond the models used for RET and OPC, computational lithographics attempts to improve chip manufacturability and yields such as by using the signature of the scanner to help improve accuracy of the OPC model:[9]

  • polarization characteristics on of the lens pupil
  • Jones matrix of the stepper lens
  • optical parameters of the photoresist stack
  • diffusion through the photoresist
  • stepper illumination control variables

A CPU-century worth of calculations or more[edit]

The computational effort behind these methods is immense. According to one estimate, the calculations required to adjust OPC geometries to take into account variations to focus and exposure for a state-of-the-art integrated circuit will take approximately 100 CPU-years of computer time.[10] This does not include modeling the 3D polarization of the light source or any of the several other systems that need to be modeled in production computational photolithographic mask making flows. Brion Technologies, a subsidiary of ASML, the largest manufacturer of photolithography systems, markets a rack-mounted hardware accelerator dedicated for use in making computational lithographic calculations — a mask-making shop can purchase a large number of their systems to run in parallel. Others have claimed significant acceleration using re-purposed off-the-shelf graphics cards for their high parallel throughput.[11]

References[edit]

  1. ^ a b "Reticle enhancement technology will extend life of 193nm litho", Electronics Weekly, 2004-02-25 
  2. ^ "Major US Semiconductor Maker Chooses TMA for OPC Software", PRNewswire, 1997-10-16 
  3. ^ McGrath, Dylan (2005-12-16), "DFM pumps up the volume", EETimes 
  4. ^ McGrath, Dylan (2005-02-12), "Litho simulation vendor opens Japanese subsidiary", EETimes 
  5. ^ LaPedus, Mark (2008-09-17), "IBM rolls 'computational scaling' for litho at 22-nm", EETimes 
  6. ^ E. Lam; A. Wong (2009), "Computation lithography: virtual reality and virtual virtuality", Optics Express 17 (15): 12259–12268, doi:10.1364/OE.17.012259, PMID 19654627 
  7. ^ A. Wong (2001), Resolution enhancement techniques in optical lithography, SPIE Press 
  8. ^ S. Chan; A. Wong, and E. Lam (2008), "Initialization for robust inverse synthesis of phase-shifting masks in optical projection lithography", Optics Express 16 (19): 14746–14760, doi:10.1364/OE.16.014746 
  9. ^ Hand, Aaron (November 2006), "Nikon and Synopsys Deliver on Advanced OPC Promise", Semiconductor International 
  10. ^ Wiley, Jim (May 2006), "Future challenges in computational lithography", Solid State Technology 
  11. ^ LaPedus, Mark (2008-02-28), "Gauda claims OPC acceleration breakthrough.", EE Times