Using oxygen, rather than air, in coal-fired boilers has been studied for several years as a strategy to reduce NOx and concentrate CO2 for capture. In combination with flue gas recirculation, higher levels of CO2 are expected but increased H2O and SO2 levels also may occur. In order to understand the role of substrate composition on corrosion, a combination of commercial and model alloys were investigated with synthetic coal ash and gas compositions simulating air- and oxyfiring environments. Exposure temperatures ranged from 600°-800°C to cover current operating temperatures up to advanced ultrasupercritical conditions. Using 500h exposures, no consistent negative effect was found for switching to the oxy-firing environment with the same synthetic ash. For model Fe-Cr alloys, 30%Cr was needed to form a thin protective reaction product across this temperature range. Among the commercial stainless steels, 310-type stainless steel showed low reaction rates with the maximum attack at 650°C. At higher temperatures, the depth of attack on Fe-base type 310 stainless steel was less than for Ni-base alloy 740. Initially, this difference was attributed to the Al and Ti additions in alloy 740. However, cast and hot rolled model Ni-18Cr and -22Cr alloys with various Al and Ti additions showed decreased metal loss with increasing Al and Ti additions in the oxy-firing environment at 700° and 800°C. As expected, metal loss was very sensitive to Cr content. A second set of model alloys also examined the effect of Co and Mo.

This content is only available as a PDF.