Jump to main content.


MODTOMON Full Paper

Information provided for informational purposes onlyNote: EPA no longer updates this information, but it may be useful as a reference or resource.

Table of Contents

I.  Introduction

II.  Raw Materials
    A. List of Pollutants
    B. Monitoring Data
        i) Completeness check and computation of annual averages
        ii) MDL
        iii) Flagging
            a) Dubious accuracy
            b) Border sites
            c) Known missing sources
    C. Emissions Inventory
        i) NTI
        ii) EMS-HAP
    D. Model Results
        i) Estimates at monitoring sites
        ii) Estimates at census tract centroids

III.  Model-to-Monitor Comparison Analysis Methods
    A. Graphical
        i) Scatter plots
        ii) Ratio box plots
    B. Statistical
        i) Number of sites
        ii) Median of ratios
        iii) Percent of sites estimated "within a factor of x"
        iv) MAXTOMON

IV.  Uncertainties
    A. History Of Model-to-Monitor Comparisons With ASPEN
    B. Uncertainties Affecting Our Comparison
        i) Monitoring uncertainties
           a) Temporal gaps
           b) Spatial gaps
           c) MDL
           d) Monitor siting
           e) Location
        ii) Emissions uncertainties
           a) Point sources
            1) Location
            2) Stack parameters, fugitive vs. stack
           b) Spatial and temporal allocation in EMS-HAP
        iii) Model uncertainties
           a) Interpolation
           b) Deposition

V.  General Results
    A. Overall
    B. Benzene
    C. Other Gases
    D. Metals
        i) Lead
           a) MAXTOMON results
           b) Location uncertainty results
             1) The Missouri source
             2) The Tennessee source
             3) The Florida source
        ii) Cadmium
        iii) Chromium

VI.  Conclusions

VII.  References

 


I. Introduction

One way to evaluate the usefulness and limitations of a dispersion model like the Assessment System for Population Exposure Nationwide (ASPEN) is to compare its ambient concentration predictions to available concurrent monitoring data.  As part of the National Air Toxics Assessment (NATA) initial national-scale assessment, the ASPEN model has predicted annual average ambient concentrations for 34 hazardous air pollutants (HAPs), 33 urban HAPs plus diesel particulate matter, at approximately 60,000 census tract locations nationwide for the year 1996.  In 1996 there were only several hundred air toxics monitoring sites all across the US.  Many of these sites, which were primarily designed and maintained under existing criteria air pollutant monitoring programs (e.g., Photochemical Assessment Monitoring Stations, or PAMS; State and Local Air Monitoring Stations, or SLAMS; Interagency Monitoring of Protected Visual Environments, or IMPROVE), only monitored a handful of HAPs for a limited period of time.  Where data allow us, we can look at the 1996 annual averages from these sites for whichever HAPs they monitor, and compare these annual averages to the 1996 annual averages generated by ASPEN for the appropriate geographic location.  By comparing the ASPEN predictions with the available monitoring data, we hope to gain a better understanding of the overall performance and limitations of the quantitative ASPEN model predictions.  It is these ASPEN model predictions that will be used in the initial national-scale assessment to predict exposure and risk values nationwide.  These predicted exposure and risk levels will subsequently help the Agency in setting priorities for future control efforts of our air toxic programs.

This document describes the results of a model-to-monitor comparison we conducted for a subset of the 33 urban HAPs.  We view this comparison as a evaluation of not only ASPEN, but also its inputs: emissions data from various sources including the National Toxics Inventory (NTI), the Emissions Modeling System for Hazardous Air Pollutants (EMS-HAP), and meteorological data.

For most of the pollutants examined, we found that the model's estimates tend to be lower than the monitor averages at the exact locations of the monitors.  However, for some HAPs, there are usually high modeled concentrations in the vicinity of the monitor.  So it may be presumptuous to say that the modeling system underestimates the monitors.  It may just be finding a peak concentration in a different place from the monitor.

However, in general, it appears that the modeling system is underestimating the monitors.  We discuss some possible reasons for this.  The National Toxics Inventory (NTI) is missing some emissions sources, and for many of the sources in the NTI, some of the emissions parameters are defaulted or missing.  It is also possible that emissions rates are being underestimated in many places.  We believe the ASPEN model itself is contributing in only a minor way to the underestimation.  We think this because output from the antecedents of the ASPEN model compared favorably to monitoring data in cases where the emissions and meteorology were accurately characterized and the monitors took more frequent readings; and in simulations we ran, the ASPEN model's estimates compared favorably to the estimates derived from a more meticulous model.  Monitor siting may have contributed to the underestimation, in that they are normally sited to find peak pollutant concentrations, which implies that errors in the characterization of sources would tend to make the model underestimate the monitors.  Finally, we are not sure of the veracity of the monitor averages, which have their own sources of uncertainty.

Our results suggest that the model estimates are uncertain on a local scale; i.e., at the census tract level.  We believe that the model estimates are more reliably interpreted as being a value likely to be found within 30 km of the census tract location.

These results are consistent with the results of an evaluation of 1990 ASPEN output,1 which was conducted as part of the Cumulative Exposure Project (CEP).  There are many differences between the evaluation for the 1990 assessment and the 1996 assessment–different evaluation methods, different emissions inventory, different list of pollutants assessed–but most of the model estimates there, as here, were lower than the monitor averages.

Because of all the uncertainties involved, we focus a good portion of our energies in this document on assessing the uncertainties.

II. Raw Materials

A. List Of Pollutants.

The EPA has conducted this comparison for seven pollutants: benzene, perchloroethylene, formaldehyde, acetaldehyde, cadmium, chromium and lead.  These pollutants were chosen because:
  1. they are a subset of the 34 air toxics considered in the national-scale assessment;
  2. they represent a range of pollutant types (i.e., organic, volatile, particulate matter);
  3. they are influenced by different combinations of mobile, area and point sources, and include reactive and non-reactive compounds; and
  4. EPA considers the available 1996 monitoring data for these pollutants to be adequate for a national-scale evaluation.
Benzene and perchloroethylene are volatile organic compounds (VOCs); formaldehyde and acetaldehyde are aldehydes; and cadmium, chromium, and lead are metals.  The VOCs and aldehydes are gaseous, while the metals are particulate matter.

B. Monitoring Data.

EPA's Office of Air Quality Planning and Standards (OAQPS) currently maintains an Air Toxics Data Archive of toxics monitoring data.  Much of these data are already publicly available via the Aerometric Information Retrieval System (AIRS).  The data not already in AIRS have been collected from various monitoring agencies by OAQPS over the past few years.  More details on the structure and contents of the Archive are available from EPA.2  EPA is hoping to make all the data in the Archive conveniently available to the public within the next year or two.

i) Completeness check and computation of annual averages.

The model estimates are annual averages for 1996, so all our monitoring data are also annual averages for 1996.  The annual averages were computed as follows:

1.  A measurement was considered below the method detection limit (MDL) if either it is indicated as being below the MDL (e.g., value of –888 in the data archive), or if it was specified with a numeric value that is lower than the reported MDL value for that pollutant/monitor/time combination.  If there was no reported MDL, the lowest reported value for the pollutant/monitor combination was assumed to be the “plausible MDL”.  Measurements below the actual or plausible MDL were assigned one of three values depending on the averaging statistic being calculated:

  1. One-half the value of the actual or plausible MDL
  2. The value of the actual or plausible MDL
  3. Zero
All annual averages used in this document are of the first type.

2.  For each pollutant/monitor combination, an annual average was calculated stepwise from temporal averages of shorter durations, as indicated below.  At each step the data set was assessed for completeness and retained for further processing only if the completeness criteria were met for the given averaging period.

    a.  Daily average.  A day was considered complete if the total number of hours monitored for that day is 18 or more (i.e., 75 % of 24 hours).  For example, 18 hourly averages, 3 six-hour averages, or 3 eight-hour averages would satisfy the daily completeness criteria.

    b.  Quarterly average.  Calendar quarters are Winter (Jan-March), Spring (April-June), Summer (July-Sept), and Fall (Oct-Dec).  A calendar quarter was considered complete if it has 75 % or more complete days out of the expected number of daily samples for that quarter, and if there were at least 5 complete days in the quarter.  To determine the expected number of daily samples, the most frequently occurring sampling interval (days from one sample to the next sample) was used; in cases of ties, the minimum sampling interval was applied.

    c. Seasonal average.  The seasons are composed of 2 quarters: Winter/Fall and Spring/Summer.  A season was considered complete if it had at least 1 complete quarter.

    d. Annual average.  An annual data set was considered complete if it had 2 complete seasons.

ii) MDL.

The MDL is the lowest level at which we have confidence in a monitored value.  It is defined in the Code of Federal Regulations3 as the lowest value at which we can be 99% confident that the true concentration is nonzero.  This value varies by pollutant and by monitor.  As described above, in the computation of the monitor averages used in this document, hourly and daily monitor readings below the MDL were replaced by one-half the MDL.  For some pollutant/monitor combinations with complete data for 1996, many of the monitor readings are below the MDL.

For these low concentration sites, we do not have much confidence in the monitored annual averages, because of the uncertainty introduced when replacing values below MDL with one-half the MDL.  Because of this uncertainty, only pollutant/monitor combinations with at least 50% of data above the MDL were included in the comparison.  In the CEP model-to-monitor comparison,1 only pollutant/monitor combinations with at least 90% of data above the MDL were included.  The cutoff percentage we chose was a compromise between ensuring high quality monitoring data and ensuring that our sample size was large enough to compute meaningful statistics.  In general, we believe that monitored averages for pollutant/monitor combinations which have more than half their observations below the MDL are too uncertain to be used in a model-to-monitor comparison.

iii) Flagging.

We discarded pollutant/monitor combinations from the model-to-monitor comparison if any of the following were true:
  • We have reason to doubt the accuracy of the monitoring average.
  • The site is very close to an international border.
  • We know that a source in the vicinity of the monitoring site is missing from the emissions inventory.
a) Dubious Accuracy.  We discarded pollutant/monitor combinations for which we had reason to doubt the accuracy of the monitoring annual average.  In areas with few major point sources, we would expect both toluene and benzene (both HAPs emitted primarily by mobile sources) to be highly correlated in the ambient air.  Thus, if a site had an unusual ratio of toluene to benzene, we discarded the site from the site list for benzene.  The same is true of formaldehyde and acetaldehyde; sites with odd ratios of formaldehyde to acetaldehyde were eliminated from the site list for both pollutants.  Sites which had extremely high monitoring values compared to other values for that pollutant were also discarded.

b) Border Sites.  Some of the monitoring sites are very close to the US-Mexico border or the US-Canada border.  Because we are not using emissions data for Mexico or Canada, we do not have confidence in the model estimates for these sites.  This is especially true of sites in Calexico, CA; El Paso, TX; Brownsville, TX; and Bellingham, WA.  These sites are close to large cities on the other side of the US border.

Even though we discarded these sites from the comparison, it should be noted that we are using model estimates at census tracts along the border in NATA.  In preliminary runs, the model estimates were much lower than the monitor averages at these sites for all pollutants.  The model estimates will almost certainly be low for census tracts near international borders, especially when large sources are located on the other side of the border.  The absence of emissions data from Mexico and Canada is a weakness of NATA.

c) Known Missing Sources.  In a few cases, by the end of the study, we discovered that a source near a monitor had been missing from the initial emissions inventory.  We discarded these pollutant/monitor combinations because they skewed our results (the model estimate is near zero, but the monitor average is significant) and because they do not provide a meaningful test of model performance.  We will add these missing sources to the NTI for the future.  Only a few pollutant/monitor combinations were discarded for this reason.

Future evaluations of the ASPEN modeling system should not exclude sites along the border or sites associated with missing emissions sources, in order to better evaluate the performance of the modeling system in its entirety.  We recommend including these sites in the analyses and explaining any disagreement between model estimates and monitor averages, instead of discarding them.

C. Emissions Inventory.

The ASPEN model uses emissions and meteorology as inputs.  The "raw" emissions inventory data is primarily from the National Toxics Inventory (NTI).  The data in the NTI are further processed and made "model-ready" by the Emissions Modeling System for Hazardous Air Pollutants (EMS-HAP).  A user's guide on EMS-HAP will be available soon.

i) NTI.

The NTI contains air toxics emission estimates for four source types: major, area and other, onroad mobile, and nonroad mobile.  Some of these sources are point sources with specific geographic coordinates; these fall into the "major" or "area or other" category, depending on the amount of emissions.  The remaining non-point sources are summarized at the county level.  We have more extensive information about the NTI elsewhere on the NATA web site, so we will not discuss too many specific details here.  The NTI is a very important input to the model.  Estimates from the ASPEN model are highly sensitive to emission rates, locations, and release parameters such as height and velocity, in the vicinity of sources.  Much of our study of the uncertainties of the model estimates in this paper focus on the NTI.  It is an enormous undertaking to estimate emissions for each source nationwide, and the NTI understandably has some differences from reality.

ii) EMS-HAP.

For mobile and area sources, NTI estimates are at the county level.  All NTI emissions represent total annual emissions.  The ASPEN model, however, requires higher resolution both temporally and spatially.  EMS-HAP allocates the emissions from the NTI temporally and spatially, for use by the ASPEN model.

Temporally, EMS-HAP allocates the annual total emissions into eight three-hour periods within an annually averaged day.  Each day of the year is allocated the same emissions.

EMS-HAP spatially allocates non-point, county-level emissions to the census tracts within each county, as required by the ASPEN model.  EMS-HAP allocates these emissions based on surrogates such as population, land use, roadway miles, etc., depending on the source category.  For point source emissions, spatial allocation is not performed because most point sources in the NTI already have exact latitude/longitude coordinates.  When point source geographic coordinates are missing, EMS-HAP defaults them where possible.  For example, if the geographic coordinates for a point source are missing but the zip code is not, EMS-HAP assigns the lat/lon coordinates of the zip code's centroid to the source.  If the zip code is missing but the county is not, EMS-HAP assigns to the source the lat/lon coordinates of a census tract centroid chosen randomly from the county.

We will discuss some of EMS-HAP's allocation and defaulting techniques in the uncertainties section of this document.

D. Model Results.

Two types of model estimates are used in this comparison: estimates at the exact location of the monitors, and estimates at every census tract centroid in the U.S.  The first type of estimates were generated especially for the purpose of doing a model-to-monitor comparison.  The second type of estimates are an integral part of NATA; these estimates are fed into the exposure and risk assessments.

i) Estimates at monitoring sites.

We used the ASPEN model to estimate concentrations at the exact locations of the monitors, to get point-to-point comparisons.  Most of the analyses in this document use the results of this point-to-point comparison.  There is nothing different about running ASPEN for monitor locations than for census tract centroids.

ii) Estimates at census tract centroids.

All other model results on the NATA web page use the tract-level model estimates.  We use the tract-level estimates in this comparison for the MAXTOMON test described in section III.B.iv.

III. Model-to-Monitor Comparison Analysis Methods

We describe some of the analytical tools used in this document below.  All but one of the following tools deal with point-to-point comparisons–they examine whether the model's estimates agree with the monitor's estimates at the exact monitor location.  The one that does not use only a point-to-point comparison is the MAXTOMON tool, which compares the monitor average to the maximum of model estimates within a circle around the monitor. This type of test allows for more uncertainty in the locations of sources and monitors, release heights, meteorology, etc.

A. Graphical.

i) Scatter plots.

Scatter plots are a relatively straightforward graphical way to show the relationship between two variables.  We simply plot model estimates of annual averages against monitor averages.  Each ordered pair on the graph is (model, monitor), for each monitoring site for that pollutant.  For example, if we have 90 monitors for benzene, we will have 90 ordered pairs to plot.  We will also show the 2:1, 1:1, and 1:2 lines on the plots.  We consider the model estimates to be reasonable for a pollutant/monitor combination if the point falls between the 2:1 and 1:2 lines.  In modeling parlance, we call this "agreement within a factor of 2".  We think the model is performing reasonably well for a given pollutant if most of the points fall between the 2:1 and 1:2 lines.

ii) Ratio box plots.

Ratio box plots show the same data in the scatter plots, in a different way.  Each box shows the distribution of model-to-monitor ratios.  So if we have 50 monitors for perchloroethylene, we will have 50 model/monitor ratios to compute.  We then compute percentiles and the mean of these 50 numbers, and create a box plot.

The plots will show the median, 25th, and 75th percentiles of the ratios.  If the model is performing well, the box plots will be short, and centered at around 1.

We decided not to show more extreme percentiles (e.g., 10th and 90th) of the ratios because often the extreme percentiles were far from the center of the distribution.  This is an unfortunate property of ratios.  When the denominator is near zero, the ratio can be enormous.  In our study, sometimes model estimates of near zero are paired with significant monitor averages, due to missing sources or sources with defaulted or incorrect locations.  These result in ratios near zero, which skew the distribution of the ratios.  Less often, we'll see high model estimates paired with low monitor averages.  But because ratios take extreme values so often, we decided to only show the interquartile range of the distribution.

There is an additional reason for leaving out the extreme percentiles.  Given values with large uncertainty, high percentiles tend to be biased high and low values tend to be biased low.4  This is because the values with large positive errors will collect in the high end of the distribution, and values with large negative errors will collect in the low end of the distribution.  The ratios are fraught with uncertainties of all kinds.

We decided to use a logarithmic scale for the vertical axis, for the following reason.  If we use a regular arithmetic scale on the vertical axis, then a ratio of 2 is twice as far from 1 as a ratio of 1/2.  But as said before, modelers typically speak of estimates as "within a factor of x."  An underestimate by a factor of x should look just as erroneous as an overestimate by a factor of x.  The logarithmic scale makes the overestimation and underestimation the same distance from the horizontal line where the ratio is 1.

The ratio box plots in this document will be shown side-by-side, one for each pollutant.  This will allow us to see easily which HAPs are being overestimated and underestimated, and which are being estimated consistently and inconsistently.

B. Statistical.

i) Number of sites.

The number of sites is the number of monitors for each pollutant which were not filtered out by any of the criteria in section II.B.  The more monitors, the more data we have, and the more we can trust the findings in the comparison.  Lead and benzene are the pollutants in this paper with the most monitors.

ii) Median of ratios.

The median of ratios is based on the model/monitor ratios for a given pollutant.  A median close to 1 suggests that the model overestimates the monitors about as often as it underestimates the monitors.  This statistic is also shown on the ratio box plots.

iii) Percent of sites estimated "within a factor of x".

This statistic is also based on the model/monitor ratios for a given pollutant.  We will often look at the percent of sites for a given pollutant which agree within a factor of 2, which is the percent of sites for which the model estimate is somewhere between half and double the monitor average.  We'll also talk about the percent of sites estimated within 30%: this is the percent of sites for which the model/monitor ratio is between 0.7 and 1.3.

iv) MAXTOMON.

This technique compares the MAXimum model estimate within r kilometers of the monitor TO the MONitor average.  All model estimates are considered in finding the maximum–both estimates at monitor sites and estimates at census tract centroids (see section II.D above).  This is an example of a point-to-range tool.  We use this tool to test whether the frequent underestimation by the model at monitoring sites was due to location uncertainties, or due to systematic underestimation.  To explain further, let's say the model estimate for a certain pollutant/monitor combination was much lower than the monitor average.  We might wonder if the model predicted a concentration similar to the monitor average anywhere near the monitor site.  If it did, it is possible that the underestimation at the exact monitor site was due to uncertainties in the inputs to the model, especially in source locations, instead of due to systematic underestimation by the model.

In general, we hope to see very few monitor averages underestimated by the model as r gets large.  One weakness of the MAXTOMON test is that there is a sparser network of model receptors in rural areas than in urban areas, because the census tracts are larger.  Thus, the MAXTOMON test might have more difficulty finding a peak concentration in a rural area than an urban area.

IV. Uncertainties

In this section, we will look at some of the sources of uncertainties which factor into the comparison.  We will begin with a historical overview of model-to-monitor comparisons done with ASPEN, to get an idea how ASPEN results compare to monitored concentrations when the comparison is done on a smaller scale, and to provide a historical context for our own comparison.  In these cases, the emissions, meteorological, and monitoring data are more likely to be of higher quality than ours.  We will then assess some of the uncertainties involved with the emissions and monitoring data, as well as some of the uncertainties introduced by the ASPEN dispersion model itself.

A. History Of Model-to-Monitor Comparisons With ASPEN.

In the early years of air dispersion modeling, say prior to 1968, calculations were completed with paper, pencil and hand calculators. Early computers were limited in their memory capabilities.  This spawned the development of a particular type of dispersion model which employed a statistical summary of meteorological conditions, which then required a special algorithm for characterizing the resulting dispersion.  Early examples of this type of model were described by Meade and Pasquill5 and Lucas.6  The idea was relatively simple, but most of the algorithms for characterizing the basic processes (e.g., buoyant plume rise, plume dispersion, depletion, etc.) were simplistic with little experimental verification.  Basically, a computation was made for each expected wind speed and stability condition whose probability of occurrence was computed for wind sectors surrounding the source (varying from 12 to 16 wind sectors).  The average concentration was computed by summing for each wind sector the computed concentration at each downwind distance, multiplied by the frequency of occurrence of each wind speed and stability combination.

Pooler7 described one of the first attempts to employ numerical methods for automating the computations (an IBM 650 computer) to provide estimates of monthly average concentration values for comparison with observations of sulphur dioxide (SO2) collected daily from November 1958 through March 1959 at 123 sampling sites in Nashville, TN.  We have to temper the evaluation results, as regressions were performed with the observed concentration values to provide best estimates of the variation of the monthly emission rates from the known sources.  When we digitized Pooler's data for reanalysis we only found 122 data values, not 123.  That said, the model overestimated the observed values by a factor of 1.37 with a correlation coefficient (r2) of 0.95.  (Note: In all descriptions of past model-to-monitor comparisons in this section, the reported factor of over or under estimation and correlation coefficient (r2) were deduced through a linear regression with the intercept forced to be at the origin.)  We found 110 of the 122 values within a factor of 2 of the observed values, with 74 of the 122 values within 30%.

Martin8 summarized the dispersion model used to numerically compute (an IBM 1130) winter season estimates of the average SO2 concentration values for comparisons with observations collected daily from December 1964 through February 1965 at 40 sites in the St. Louis area.  Removing comparison results for the five most suspect locations reduced the overestimation to 1.47 and increased the correlation coefficient to 0.95.  Thirty-four of the 35 values were within a factor of 2, with 14 within 30%.  A reanalysis of these same data was performed by Calder9 using the Climatological Dispersion (CDM) model, using a revised characterization of the area source emissions by Turner and Edmisten.10  A major enhancement within the CDM over the model employed by Martin was to incorporate an improved treatment for characterizing dispersion from area sources, employing an algorithm based on the narrow plume hypothesis.  In spite of the attempts to improve the characterization of area source emissions and the dispersion from these low-level sources, the comparison results were similar to those achieved by Martin.  The CDM tended to overpredict concentration values by a factor 1.54 with a correlation coefficient (r2) equal to 0.92.  Thirty-five of 39 estimates where within a factor of 2, with 17 within 30%.  Possible factors contributing to the tendency to overestimate the observed concentration values were: an inherently crude emissions inventory, no day versus night variation in emission rates, and the crude estimates of mixing height employed.

Turner et al.11 summarized the results obtained in applying the CDM model to estimate annual average particulate and SO2 concentration values for the New York area for 1969.  SO2 observations were available for comparison at 75 locations and total suspended particulate matter observations were available for comparison at 113 locations.  This version of the CDM employed the Briggs12 plume rise algorithms (in contrast to use of Holland13 algorithms used by Martin and Calder in the St. Louis comparisons).  For SO2 it appears the CDM tended to slightly overpredict concentration values by a factor of 1.11 with a correlation coefficient (r2) equal to 0.90.  Seventy-one of the 75 values were within a factor of 2, with 47 values within 30%.  For particulates it appears the CDM tended to slightly underpredict concentration values by a factor of 0.93 with a correlation coefficient (r2) equal to 0.94.  111 of the 113 values were within a factor of 2, with 94 within 30%.

Irwin and Brown14 summarized the results obtained in applying the CDM model to estimate 1976 annual average SO2 concentration values for the St. Louis area.  There were 13 sites, but omission of a lead smelter near one site precluded use of data at two sites for model performance comparisons.  The emission inventory and monitoring results were obtained as part of the St. Louis Regional Air Pollution Study.  These simulations differ with those computed by Turner et al. in that urban dispersion parameters (based on tracer studies conducted in St. Louis, McElroy and Pooler15 and Gifford.16  It was determined that although the area source emissions constituted only 3.5% of the total area and point source emissions, estimated concentrations from area sources ranged from 14 to 67% of the total concentration estimated at the monitoring sites.  For the 11 sites it was found that CDM slightly overpredicted concentration values by a factor of 1.10 with a correlation coefficient (r2) equal to 0.96.  Nine of the 11 sites have estimates within a factor of 2, with 3 values within 30% of those observed.  This same inventory was simulated using the RAM model,17 which employed hourly specification of the meteorology and the emissions.  For the 11 sites it was found that RAM slightly overpredicted concentration values by a factor of 1.10 with a correlation coefficient (r2) 0.96.  For the RAM estimates, all 11 sites had estimates within a factor of 2, with 10 values within 30% of those observed.

The version of CDM applied by Irwin and Brown is similar to the Industrial Source Complex (ISCLT) Long-Term model.18  The major difference in ISCLT versus CDM is that the area source algorithm is better than that employed in CDM.  ISCLT's area source dispersion characterization nearly approximates what is obtained when one computes area source impacts using an hour-by-hour simulation (which employs a double integral over the area and hence is our best expression of dispersion from an area).  The emphasis on improving the treatment of area source impacts reflects the recognition that area source emissions (if present) often account for a major portion of the simulated impacts, as discussed in the previous paragraph.

In the studies summarized in the table below, it is important to remember that the long-term models have evolved, with the adoption of improved characterizations for the dispersion and for treatment of area sources.  Except for the simulations for Nashville by Pooler and for St. Louis by Martin and Calder, the average bias has been slight.  The CDM versus RAM comparisons offer an interesting clue that perhaps the time variation of the emission rates (which Calder was the first to offer as a major concern) is of importance, although there are other differences between CDM and RAM that also might offer explanation of the differences seen.  Regardless of the model employed, the estimates are generally within a factor of 2 of those observed.


Table 1.  Summary of long-term model simulation comparisons.  Note the tendency for over or underprediction.  The correlation coefficient (r2) is based on a linear regression with the intercept specified as the origin.
Study Number of
Values
Over/Under 
Factor
Correlation
(r2)
Within Factor
of 2
Within 30%
Pooler (1961) Nashville SO2 123 1.37 0.92 90% 61%
Martin (1971) St. Louis SO2 35 1.47 0.94 97% 40%
Calder (1971)
St. Louis SO2 CDM
39 1.54 0.92 90% 44%
Turner et al. (1971)
New York SO2 CDM
75 1.11 0.90 95% 49%
Turner et al. (1971)
New York Particulates CDM
113 0.93 0.94 98% 83%
Irwin and Brown (1984)
St. Louis SO2 CDM
11 1.10 0.96 82% 27%
Turner and Irwin (1983)
St. Louis SO2 RAM19
11 1.10 0.96 100% 91%

B. Uncertainties Affecting Our Comparison.

i) Monitoring uncertainties.

It is tempting to treat the monitoring data as "truth": a fixed target which the model is trying to hit.  However, our monitoring data is a coarse conglomeration of any monitoring data available.  In more small-scale, careful model-to-monitor comparisons, such as the ones considered in the previous section, it is more likely that only high-quality monitoring data is used.  But our comparison is conducted on a large scale, with data from a wide variety of monitoring agencies.  Unlike the criteria air pollutant world, there currently is no formal national air toxics monitoring network which follows standardized EPA guidelines or established national monitoring procedures.  While several States and local agencies have collected some high quality HAP monitoring data, some of the data have not undergone any formal quality assurance tests, and the data come from several different monitoring networks, which may vary in precision and accuracy.  In general, we would expect the precision and accuracy of air toxics monitoring data to be not nearly as good as the SO2 and particulate matter monitoring data used in the studies in the previous section.  We will discuss some of the other monitoring uncertainties in more detail below.

a) Temporal gaps.  Most of the 1996 data in the Archive were not collected every day.  Instead, they were produced every 12th day or every 6th day throughout the calendar year.  Our completeness criteria discussed in section II.B.i above filter out pollutant/monitor combinations which have many missing sample days, but they do not filter out pollutant/monitor combinations based on the number of missing calendar days.  Our only requirement for the calendar days is that we have at least five days in either the winter or fall quarter and in either the spring or summer quarter.  This is not a very rigorous standard.  The temporal gaps are especially problematic for pollutant/monitor combinations which have widely varying concentrations throughout the year.  This might be true for pollutant/monitor combinations where the weekday and weekend concentrations are very different, or those near sources whose emissions vary widely from day to day.

b) Spatial gaps.  For some of the pollutants in this study (section II.A above), we have very few sites across the country.  The table below shows the number of monitors and the number of states covered for each pollutant.

Table 2.  Geographic coverage of monitoring data, by pollutant.
Pollutant Number of Sites Number of States
Benzene 87 16
Perchloroethylene 44 8
Formaldehyde 32 10
Acetaldehyde 32 10
Lead 242 28
Cadmium 20 7
Chromium 36 6

We only have 20 complete monitors for cadmium, 13 of which fall in Illinois and New York.  Acetaldehyde and formaldehyde have exactly the same set of 32 monitors, 28 of which fall in the northeastern and Great Lakes states.  Of the 36 chromium monitors, 32 are in California, New York, and Illinois.  All 10 chromium monitors in New York are in Staten Island.  The situation for benzene and lead is much better, as the table shows.  In general, the southeastern, northwestern, Great Plains, and Rocky Mountain states are very sparsely monitored.  The northeastern and mid-Atlantic states, Great Lakes states, California, Texas, and Louisiana are fairly well monitored.

c) MDL.  As discussed in section II.B.ii above, values below MDL add uncertainty to the annual averages.  We dealt with this by eliminating from the comparison all pollutant/monitor combinations for which less than 50% of the daily observations were above the MDL.  Still, many of the pollutant/monitor combinations have barely more than 50% of daily observations above the MDL.  Of the 493 pollutant/monitor combinations, 36 (7.3%) have between 50% and 60% of daily observations above the MDL.  Chromium is the most uncertain pollutant in this respect: 13 of the 36 monitors (36%) have less than 60% above the MDL.


 

Table 3.  Percent of daily values above MDL, by pollutant.  Many of the monitors have a large percentage of data below MDL, especially for chromium.
Number of Sites With Percent of Daily Values Above MDL
Pollutant Number of Monitors 50% to  <60% 60% to <70% 70% to <80% >=80%
Benzene
87
8
2
4
73
Perchloroethylene
44
2
0
1
41
Formaldehyde
32
0
0
0
32
Acetaldehyde
32
0
1
0
31
Lead
242
12
8
11
211
Cadmium
20
1
1
1
17
Chromium
36
13
3
4
16
All
493
36
15
21
421
EPA is currently discussing ways to handle values below the MDL other than just replacing them with MDL/2.  Other methods may reduce the uncertainty we have in monitor averages when many daily observations are below the MDL.

d) Monitor siting.  The monitors used in the study were located considering a wide variety of objectives.  Some were placed near sources, to monitor sites with high pollutant concentrations.  Others were located in residential areas, intended to measure more typical exposures.  Still others were located in rural areas, in order to find a background concentration.  A few were run for special studies.  The Archive has information on siting objectives for some of the monitors, but not all: 179 (36%) of the 493 pollutant/monitor combinations used in this comparison had missing site objective codes.  Also, there are no specific guidelines for when an air toxics monitor should be classified as "source-oriented", "population-oriented", "background-oriented", etc., so even for those monitors with siting objective information in the Archive, we cannot be sure of the quality or consistency of such data.
    Generally, most monitors are sited to find peak concentrations.  We use this assumption in the interpretation of the MAXTOMON results.

e) Location.  How trustworthy are the exact (latitude,longitude) coordinates of the monitors used in this comparison?  This is difficult to assess.  One way to assess this is to see how often the monitor's lat/lon coordinates fall in a different county from the one reported in the Archive.  Another is to take a Global Positioning System (GPS) out to some monitors and check its readings with the values in the Archive.
    We performed the first test using a Geographic Information System (GIS) software package and the US counties coverage which is included with the software.  Of the 493 pollutant/monitor combinations, only one has lat/lon coordinates which fall more than 5 km from the reported county.  This monitor is in Louisiana.  We didn't look at monitors which are outside of the reported county by less than 5 km, because these discrepancies are likely due to lack of geographic detail in the counties coverage rather than erroneous lat/lon coordinates.  The one in Louisiana only missed by 7.5 km, so even that one might be due to lack of detail in the coverage, rather than erroneous lat/lon coordinates.
    We performed the second test using three lead monitors near a source in Herculaneum, Missouri.  Mick Daye, an EPA modeler from Region 7, took GPS readings at each monitor.  We compared Mick's coordinates to those in the Archive, and here is what we found:


Table 4.  Comparison of lat/lon coordinates for three monitors in Archive to GPS readings.  These discrepancies may affect model-to-monitor agreement.
AIRS Monitor ID Monitor Name Distance Apart (meters)
290990015 Broad Street 80
290990005 High School 50
290990011 Bluff 170

These are significant differences for a model-to-monitor comparison, which we will see in the emissions uncertainties section.  Since most lead emissions come from widely spaced industrial plants, with little contribution from well dispersed sources, there are often steep concentration gradients near a source.  A difference of 50 meters might not have much of an effect, but a difference of 170 meters would.  We will discuss this further in the emissions uncertainties section: it is very important to get the source-receptor geometry correct when using a model to estimate concentrations near an isolated point source.

ii) Emissions uncertainties.

In light of the discussion in section IV.A above, we believe that the accuracy of the emissions, including source location data, is the single most important factor affecting the performance of the model.  Thus, we focused a lot of our attention on the uncertainties surrounding the emissions.  Missing point sources is a major problem when it occurs, and difficult to assess.  The majority of the emissions inventory was compiled in cooperation with state and local agencies.  We considered emissions data submitted from these agencies to be the best data sources in the enissions inventory.  Where agencies did not submit data and where there were obvious omissions, we supplemented with emissions information gathered via EPA regulatory development and with the industry-supplied emissions reported to EPA Toxics Release Inventory (TRI).  One convenient way we have to assess whether a state is likely to have missing sources is its level of participation in submitting a point source inventory to EPA.

a) Point sources.  Of the pollutants we investigated, the point source contribution to the modeled concentrations is highest for the three metals.  Thus, our investigation of point source uncertainties will focus on the metals.  Some of the emissions data variables that are important for accurate modeling are location, release height, and emissions rate.  We discuss uncertainties surrounding the first two variables below.

1) Location.  We will now give a brief explanation of the way EMS-HAP processes the point source location data from the NTI.   It does this in two major ways.  First, if the geographic coordinates of the source are missing or out of range (i.e., outside the U.S., Puerto Rico, and Virgin Islands), it places the source at the centroid of the zip code if that is nonmissing.  If the zip code is also missing, it places the source at a census tract centroid chosen randomly from the county.  Second, for the sources with nonmissing geographic coordinates, it checks if the lat/lon coordinates fall within the reported county.  If not, but the coordinates are within a distance of 5.4 times the county radius (the county is approximated as a circle of area equal to the county area) from the county centroid, EMS-HAP leaves the lat/lon coordinates as is.  If the coordinates are outside this circle, EMS-HAP will move the source to the zip code centroid if available.  If not available, it will move the point to a randomly selected census tract centroid in the reported county.  For more detail, please consult the EMS-HAP user's guide.

There are two ways we have assessed the location uncertainties of the point sources used in ASPEN.  First, we found out which source lat/lon coordinates fall more than 5 km outside their reported counties, just as we did for the monitor locations.  Second, we found out which sources have locations that were defaulted based on either the zip code or county defaulting methods.  We did this for each of the three metals.

In the table below, we looked at the percent of emissions (by mass) from each state/metal combination coming from uncertain locations.  Each cell in the table is a link to a pie chart, which shows in more detail the breakdown of emissions into categories of location uncertainty.  We also noted the states which we feel are likely to have missing sources.  As said above, some states were more forthcoming than others with respect to submitting a point source inventory.  The less forthcoming states are considered more likely to have sources not included in the NTI.  Missing sources often cause drastic underestimates by the modeling system.


Table 5.  Uncertainty in point source location data for metals, by state.  Starred states were not as forthcoming in submitting point source inventories to EPA.  State/metal combinations are denoted by O, X, XX, or XXX, depending on the percent of point source emissions by mass which have either defaulted locations or locations falling outside the reported county.  Each state/metal combination also has a pie chart associated with it, showing the breakdown of point source emissions into the location categories.
State Lead Cadmium Chromium
Alabama*
XXX
XXX
X
Arizona
XX
O
O
Arkansas
X
O
O
California
O
XX
X
Colorado
O
O
O
Connecticut*
XX
X
X
Delaware
O
O
O
District of Columbia
O
O
O
Florida
O
XXX
O
Georgia*
XXX
XXX
O
Idaho
O
XXX
XXX
Illinois
O
O
O
Indiana
XXX
XX
XXX
Iowa*
O
O
O
Kansas
O
O
O
Kentucky
X
O
O
Louisiana
XX
O
X
Maine
O
O
O
Maryland
X
O
O
Massachusetts*
O
O
O
Michigan*
X
X
X
Minnesota
XX
XXX
X
Mississippi
XX
XXX
XX
Missouri
O
XX
X
Montana*
X
O
O
Nebraska
O
O
O
Nevada*
O
O
O
New Hampshire
O
O
O
New Jersey*
O
O
O
New Mexico
O
O
O
New York
X
XX
O
North Carolina
O
O
O
North Dakota
O
O
O
Ohio*
XX
XX
O
Oklahoma*
XX
XX
O
Oregon
XXX
XXX
X
Pennsylvania
X
XXX
O
Puerto Rico
XXX
XXX
XXX
Rhode Island
XXX
XXX
XXX
South Carolina
XX
X
X
South Dakota
O
O
O
Tennessee
X
XX
X
Texas
XX
XXX
O
Utah
O
O
O
Vermont
X
X
X
Virginia
XX
XXX
X
Washington
XX
XX
X
West Virginia
O
O
O
Wisconsin
O
O
O
Wyoming
O
O
O

All states are placed into three categories based on their level of participation in submitting a point source inventory to the NTI: high, medium, and low.  The states in the latter two categories are starred.  As for the actual entries in the cells of the table, a state/pollutant combination is assigned an 'O' if less than 10% of its emissions are "location uncertain" (either from a source with a defaulted location, or from a source falling more than 5 km outside a county boundary); an 'X' if between 10% and 25% of sources are location uncertain; an 'XX' if between 25% and 50% of sources are location uncertain; and an 'XXX' if more than 50% are uncertain.  The more location uncertainty, the less faith we have in model results for the given state/pollutant combination on a local scale.  The pie charts show the percent of emissions which fall into each of the four location categories: county defaulted; zip code defaulted; not defaulted but more than 5 km outside the reported county; and not defaulted and either inside the reported county or less than 5 km from its boundary.  All categories but the last are considered "uncertain".  The less emissions which fall into the first three categories, the more we can trust the model results on the local scale (the likelihood of missing sources should also be considered).

The table below shows the percent of emissions falling into each location category, when summing across states.  The NTI has exact geographic locations for most of the point source chromium emissions, most of which fall in the reported county; but about 13 percent of lead emissions and 25 percent of cadmium emissions were defaulted.


Table 6.  Uncertainty in point source location data for metals, for all states.
% Emissions By Mass
Metal
County Defaulted
Zip Code Defaulted
Not Defaulted, Outside County
Not Defaulted, Inside County
Lead
12.8%
0.3%
3.8%
83.1%
Cadmium
25.0%
0.2%
6.4%
68.4%
Chromium
3.5%
0.1%
3.0%
93.4%

In section V.D below, we'll discuss some of the results of the location defaulting for lead in more detail.  We did a detailed study of some of the monitoring sites which were matched up with extremely low model estimates.

2) Stack parameters, fugitive vs. stack.  The other stack parameters are as important to model results as the locations.  Slight changes in the stack parameters can cause widely varying model results.  We'll focus mainly on the release heights.  Most of our monitors measure the concentration of a pollutant on the ground; and the ASPEN model does the same.

            EPA modelers agree that the release height is important, because emissions released from high stacks have more air to pass through on their way to the ground than emissions released from ground-level.  As a result, we might expect positive errors in release height to lead to model underestimates, and vice versa.  Studies have shown that ground-level concentrations are 5 to 8 times more affected by low level emissions than by elevated emissions.20,21

In part to check the accuracy of the stack release heights in the NTI, members of EPA visited a lead smelter in Herculaneum, Missouri.  According to the TRI, 89.91 tons of lead were emitted from the 550-foot stack in the center of the facility, and 7.66 tons were "fugitive" emissions: that is, escaping from the facility through open doors, windows, etc.  However, the emissions for an ongoing 2-month study at the facility suggest that the “fugitive” emissions are of order 50 tons rather than 7.66 tons.  We can not generalize to all other sources from one site visit, but this does reveal the possibility of the types of emission characterization uncertainties that can occur.  All other factors being equal, an increase from 7.66 to 50 tons in low-level emissions would likely increase the predicted annual average for this site by a factor of 3.

b) Spatial and temporal allocation in EMS-HAP.  One of the tasks of EMS-HAP is to allocate NTI emissions summarized at the county level both temporally and spatially.  EMS-HAP makes numerous assumptions in these allocation processes, which add to the uncertainty of the model estimates.  The EMS-HAP user's guide will be available to the public soon.  This guide will describe these allocation techniques in detail.  For our purposes, it is important to note that pollutants for which area sources contribute significantly to model estimates have more uncertain model estimates.

iii) Model uncertainties.

A dispersion model in general makes many simplified assumptions as to the fate and transport of the emission plume.  One of the key simplifications of the ASPEN model is that it does not include a terrain component in its prediction algorithms.  Further, the model relies on steady-state long-term sector-averaged climate summary data to represent the conditions at any given plume site.  The model also simplifies some complex atmospheric chemical processes and captures only pollution transport within 50 km of any individual source.

Dispersion Calculations
ASPEN was constructed using Version 2 of the Industrial Source Complex Long-Term model (ISCLT2).  The major changes between Versions 2 and 3 (the current version) of ISCLT were: a new area source algorithm, a revised dry deposition algorithm, a wet deposition algorithm, COMPLEX1 algorithms, a pit retention algorithm.  None of these changes would affect ASPEN, as ASPEN does not use the ISCLT area source algorithm, dry deposition algorithm, or wet deposition algorithm.  Hence, for the simple point source case, one would expect ASPEN and ISCLT to provide similar results.  To test this, a series of runs with both ASPEN and ISCLT were made, to provide a means for comparing the respective model's estimates.

ASPEN Calculations
The ASPEN calculations are actually distributed over a series of separate processors.  The first (ASPENA) reads in the emissions data, and computes for each source the concentration for a set of receptors distributed around the source location.  Receptors are located along 16 radials outward from the source at 12 locations (ranging from 100 m to 50 kilometers).  The 16 radials are defined in a clockwise sense about the source, with the first radial pointed due north.  The interpolation procedures are all computed within ASPENB, which reads in the results from ASPENA, and combines this with a listing of where concentration values are desired.  In the simplest of situations, when no spatial averaging is employed (valid for all sources once the receptor is no longer located within the same census tract as the source), concentrations are interpolated using the values computed along the 16 radials.  Linear interpolation is used for a point between radials at a fixed distance downwind.  Log-log interpolation is used for determining concentration variations as a function of distance downwind.  We will not attempt here to describe the averaging procedures used for computing point source and area source impacts for receptor locations within the census tract of the source.  In principal, the impacts to receptors that are within the same tract as the emission are essentially computed as an area-weighted average concentration for all ASPENA receptors found within the census tract.

Test Case Meteorology
Both ASPEN and ISCLT use a Stability Array to describe the frequency of occurrence of wind speeds and stability conditions as a function of sixteen wind sectors going clockwise from the north.  For the test cases to be described, only winds from the north were specified.  Only neutral stability was allowed.  The mixing height was set at 1000 m and the annual average temperature was set at 279.65 K.  The frequency of occurrence of the six wind speed categories was specified as:


Table 7.  Frequency distribution of wind speed categories used in test cases.
Wind Speed Category 1 2 3 4 5 6
Average Wind Speed (m/s) 1.5  2.5 4.5 7.0 9.5 12.5
Frequency of Occurrence 0.315  0.169  0.290  0.197  0.029  0.000

a) Interpolation.  We specifically investigated whether the interpolation scheme used within ASPEN might be underestimating the actual modeled impacts.  This concern arose because a "net" of receptors is employed by ASPEN, and then concentrations at specific points are estimated by interpolating within the "net".  We wondered whether ASPEN might underestimate peak ambient concentrations because it "averages out" the peak values by combining them with lower concentrations nearby.
    To do this, we simulated three types of emissions sources, and compared the ASPEN estimates downwind from each source to the estimates derived from a more recent, detailed version of ASPEN, the Industrial Source Complex Long-Term Model Version 3 (ISCLT3).  The simulations were run under a variety of wind speed conditions.

The first simulated source was a point source with a 10-m stack height.  The stack gas temperature was set to be nearly the same as the ambient temperature, and the exit velocity was set at 1 m/s.  In this case, there would be no plume rise.  Figure 1 depicts the results obtained from the two models.  It is evident that at 250 m downwind, ISCLT provides a concentration that is greater than what ASPEN would provide using its log-log interpolation procedures.  However, at most distances the two models provide similar results.


Figure 1. Comparison of concentration estimated by ASP.
Figure 1.  Comparison of concentration estimated by ASPEN and ISCLT3 for a 10-m point source, with no plume rise.  The emission rate was 1 g/s.

The second simulated source is again a point source with a 10 m stack height.  In this case the plume was made buoyant, with an exit velocity of 2 m/s and an exit temperature of 495 K.  The actual plume rise (Dh) is dependent on the wind speed (u), as Dh=357/u, where Dh is in meters and u is in m/s.  Figure 2 shows the concentrations obtained from the two models for this buoyant source.  Notice that when ASPEN is run in its normal mode, where gradual rise is used at all distances, the ASPEN concentration at 100 m downwind is much larger than what ISCLT estimates.  Whereas if we run ASPEN using the final plume rise at all distances (which is how ISCLT runs), the estimates are more in line with what ISCLT provides.  The ASPEN estimates are lower than ISCLT’s by about 10% in the near distances, with the underestimation increasing to about 25% at 30 km downwind.

Figure 2. Comparison of concentration estimated by ASPEN and ISCLT3 for a 10-m point source, with plume rise.
Figure 2.  Comparison of concentration estimated by ASPEN and ISCLT3 for a 10-m point source, with plume rise.  The emission rate was 1 g/s.

As a third case, two area sources of different size were simulated within ISCLT for comparison with those estimated by ASPEN.  ASPEN does not explicitly assign a size to area sources, so we tried to deal with the two different source sizes by varying the area of the census tract.  In these comparisons, the emission rates were 1 g/s from each area source (which is expressed as g/s-m2 in ISCLT).  Figure 3 depicts the comparison results obtained.  As seen in Figure 3, once one is 3 km or more downwind, the differences are less than 20%.

Figure 3. Comparison of concentration estimated by ASPEN and ISCLT3 for area sources with release heights of 1 m.
Figure 3.  Comparison of concentration estimated by ASPEN and ISCLT3 for area sources with release heights of 1 m.  The emission rate was 1 g/s.  The area source size for the ISCLT runs was 100m by 100m and 1km by 1km.

In these comparisons, we see no systematic bias in the ASPEN calculations that would cause it to significantly underestimate concentration values in comparison to what ISCLT can provide.  If anything, it may be that ASPEN may provide higher concentration values for receptors near or within tracts with area source emissions.

b) Deposition.  The ASPEN model simulates the effect of dry deposition of particulate by adding an additional decay term to the emission rate in calculation of ambient concentrations. The decay term is a function of the deposition velocity, downwind distance from the source, and plume dimensions with respect to the mixing height.  Deposition velocity is also a function of the particle size, wind speed, and the land-use type.  The ASPEN model allows different deposition options for fine and coarse particulate and urban/rural environments.  In order to analyze the effect of these options on the modeled ambient concentrations, we performed test case simulations using lead emissions from mobile non-road sources in Colorado.  We used different compositions of fine/coarse fractions and held the total emission rate constant.  Five different scenarios were used for this test case: 10% fine and 90% coarse, 25% fine and 75% coarse, 50% fine and 50% coarse, 75% fine and 25% coarse, and 90% fine and 10% coarse.  Emissions from 17 pseudo-point sources of 10 m height, 1 m/s exit velocity, and T = 295 K were considered.  For fine particles, the deposition velocities are generally similar and scattered around the 1:1 ratio line.  The deposition velocities for coarse particles are much higher for ASPEN than for ISCST3.  The effects of these differences was extrapolated to the national scale.  For the entire U.S., the total lead emissions were 66.5 g/s and the percent contribution from different source categories was as following: 49% all lead emission were accounted for by major sources, 28% by area sources, less than 0.01% by mobile on-road, and 23% by mobile non-road sources.  For the ASPEN simulations this means that about 50% of all lead emission sources were treated as point sources and about 50% as pseudo point sources.  We estimated that ASPEN has a bias to predict average lead concentrations in the air 20–30% lower than one would typically predict, because it is using coarse particle deposition velocities that are higher than one would usually use.
    The same logic applies to the other particulate HAPs in the comparison (cadmium and chromium).  We expect that the percent of emissions in the coarse category for these other two is less than that for lead, however, so we would expect the underestimation bias also to be less.

V. General Results

A. Overall.

The ratio box plot below gives a general idea of the results of the point-to-point comparison.

Figure 4. Ratio box plot showing distribution of model/monitor ratios for each pollutant.
Figure 4.  Ratio box plot showing distribution of model/monitor ratios for each pollutant.  The bottom of each box is the 25th percentile, the top is the 75th percentile, and the horizontal line in the middle is the median.  See section III.A.ii for more details.  Also note the number of sites summarized in each box plot (see Table 8 below).

For comparison to the results from the historical studies listed in section IV.A above, the table below gives the percent of sites estimated within 30%, and within a factor of 2.  It also reports the percent of sites which are underestimated by the modeling system.


Table 8.  Agreement of model and monitors by pollutant, on a point-to-point basis.  Compare to Table 1.
Pollutant Number of Sites Median of Ratios Within Factor of 2 Within 30% >Underestimated
Benzene 87 0.92 89% 59% 59%
Perchloroethylene 44 0.52 55% 32% 86%
Formaldehyde 32 0.65 53% 28% 88%
Acetaldehyde 32 0.60 59% 22% 91%
Lead 242 0.17 18% 10% 91%
Cadmium 20 0.18 15% 5% 85%
Chromium 36 0.15 28% 19% 83%

These results are surprising given the results of the historical studies.  The agreement in our results on a point-to-point basis is comparable to the agreement from historical studies for benzene only.  The rest of the pollutants show poor agreement on a point-to-point basis, with the model estimates systematically lower than the monitor averages.  From the ratio box plot graph, we can see that this is especially true for the three metals, which all have ratio medians of less than 1/5: on "average", they are underestimated by more than a factor of 5.  This is most interesting for lead, because this is a well-studied criteria pollutant, for which we have extensive monitoring experience and an emissions inventory which has been assembled over many years.

Because the past model-to-monitor studies in section IV.A show much better agreement than ours, something must have been different in our study from historical studies–and we do not think the underestimation is due to the model itself, because the model we used here is very similar to the model used in historical studies.  We can think of two possible explanations for the underestimation:

  1. The emission rates are systematically underestimated and/or many sources are missing from the emissions inventory.
  2. Many of the monitors likely were sited to find peak concentrations.  Often, the ambient concentration falls off quickly around the peak area.  Even given a perfect model and perfect monitors, if the monitor is right at the peak and the emissions or meteorological inputs are off even a little, the model will underestimate.  This is especially likely for pollutants dominated by point sources with elevated releases, because any errors in release height, exit velocity, or emissions location might cause the model to find a peak concentration area different from the true peak.
The MAXTOMON statistic described in section III.B.iv is especially designed to investigate the second hypothesis.

We are currently designing a national Air Toxics monitoring network.  To get an idea of how to proceed with the national network, we are conducting pilot monitoring studies in a few cities across the US.  The monitors in these studies are sited in high, medium, and low concentration areas, with a future model-to-monitor comparison as one of the goals.  We believe in terms of a model-to-monitor comparison, the new network will be more useful than the current collection of monitors.

B. Benzene.

The ratio box plot (Figure 4) and short list of statistics (Table 8) in the previous section show good agreement for benzene.  Below is the scatter plot.

Figure 5. Model-to-monitor scatter plot for benzene.
Figure 5.  Model-to-monitor scatter plot for benzene.  Most points fall within the factor of two wedge, and none are far outside the wedge.

As expected from Figure 4 and Table 8, most of the points in the scatter plot fall between the 2:1 and 1:2 lines.  The high concentration monitors seem to be estimated a little less reliably: most of the points falling outside the "factor of 2 wedge" are those with high monitor concentrations.  Some of the misses are low and some are high, but nothing misses by much.  The largest model-to-monitor ratio is 2.45 and the smallest is 0.34, so all monitors are estimated within a factor of three.

There are several reasons why we would expect good agreement for benzene:

  • It is a widely distributed pollutant which is emitted from point, area, and mobile sources.  Thus, if the model is biased in the way it handles any one of these source categories, the bias will be dampened.  For example, if a point source near a monitor is mislocated away from the monitor, there will nevertheless be significant mobile source emissions near almost every monitor location which will dampen the underestimate.
  • We have an estimated background concentration for it.
  • There is a large number of monitoring sites for it (87), giving an adequate sample size for the statistics in the comparison.
  • Monitoring technology for it has a long history, suggesting that the monitoring data is reflective of actual ambient concentrations.
  • Its emissions have been tracked for a long time, so there is some confidence in emission estimates.

C. Other Gases.

In the ratio box plot in section V.A., we can see that agreement is similar for the three other gases in the study: perchloroethylene, formaldehyde, and acetaldehyde.  The model's estimates tend to be lower than the monitor averages, but the ratio medians are all within a factor of 2.

Perchloroethylene is dominated by area sources.  For the average census tract across the US, area sources are responsible for 49% of the perchloroethylene model estimate.  Modeled concentrations of both aldehydes are dominated by mobile sources, both onroad and nonroad: the percent contribution of mobile sources is 69% for formaldehyde and 90% for acetaldehyde.  For area sources and mobile sources, we are heavily dependent on the spatial allocation methods of EMS-HAP, which add uncertainty to the model estimates.

Because of the uncertainty involved in the spatial allocation methods, it is possible that the model is estimating a concentration at or higher than the monitor average nearby, but not at the actual monitor location.  Thus, let's look at the MAXTOMON results for all three gases.  We include benzene also for comparison purposes.

Table 9.  MAXTOMON table for the four gases.  The two VOCs have high modeled concentrations near the monitors most of the time, even if they are underestimated at the exact monitor locations.  This is also true for the two aldehydes, but to a lesser degree.
Percent Missing Low at Radius Of:
Pollutant # Monitors 0 km
(Exact Monitor
Location)
2 km 4 km 6 km 8 km 10 km 20 km 30 km
Perchloroethylene 44 86% 73% 61% 59% 52% 43% 23% 9%
Formaldehyde 32 88% 81% 78% 69% 59% 56% 31% 31%
Acetaldehyde 32 91% 91% 84% 69% 56% 56% 38% 34%
Benzene 87 59% 47% 36% 30% 26% 25% 20% 11%

The percent of monitors underestimated drops off quickly for perchloroethylene.  The model's estimate is low on a point-to-point basis 86% of the time; but there are modeled concentrations nearby which are greater than or equal to the monitor average for many of the monitors.  Less than half the sites are underestimated if we go out 10 km; and less than a quarter are underestimated if we go out 20 km.  This suggests that uncertainties in the location of the nearby area sources may be responsible for the underestimation on a point-to-point basis.

The effect is less dramatic for the two aldehydes, but still evident.  If we go out 20 km, 31% of the formaldehyde and 38% of the acetaldehyde monitors are underestimated, compared with 88% and 91% at the actual monitor locations.

The aldehydes differ from the five other HAPs examined in this comparison because a significant portion of their ambient concentrations were secondarily formed.  The other HAPs are relatively inert, which makes them easier to model.  However, the ASPEN model simulates atmospheric chemical reactions for the aldehydes in a simple manner.  This is an additional source of uncertainty.

It is possible that the ASPEN model underestimates the amount of secondarily formed aldehydes.  Analysis of ASPEN modeled nationwide mean values for formaldehyde and acetaldehyde suggest that 23% and 58% (respectively) of the total modeled concentrations are attributable to secondary formation.  A more recent study using OZIPR (a photochemical grid model) suggests that secondary formation generally accounted for approximately 90% of the ambient formaldehyde and acetaldehyde.22

These results loosely suggest, especially for perchloroethylene, that the model is not necessarily systematically underestimating ambient concentrations: it may just be finding the peak concentration in the wrong place.

D. Metals.

As said above, the underestimation at the actual monitoring locations is severe for the three metals, all of which have ratio medians of less than 1/5.

i) Lead.

Lead monitored concentrations in the US today tend to be high only near lead point sources.  Of the 242 lead monitors in the study, 106 (44%) are designated as source-oriented.  Both the source-oriented monitors and the other monitors are underestimated by the modeling system at the monitor locations.  Typically (using the median), the source-oriented monitors are underestimated by a factor of 7.5, and the others are underestimated by a factor of 4.9.  Only 17% of the source-oriented monitors and 18% of the other monitors are estimated within a factor of 2.

Since the underestimation at the source-oriented monitoring sites is especially severe, we investigated more closely the model estimates at some of these monitor sites.  We did not look at all 106 monitors which were labeled as source-oriented in the Archive, however.  The decision to define a monitor as source-oriented is somewhat subjective.  A monitor placed well downwind of a source in a residential neighborhood might be considered source-oriented by some but population-oriented by others.  Because of this, we chose to look closely at a subset of the monitors which we are more certain fit the mold of a source-oriented monitor, subsetting by monitored concentration.  Experience suggests that a monitor with an annual average of 0.3 mg/m3 or higher is certain to be source-oriented.

Of the 242 monitors used in the comparison, 42 are both labeled as source-oriented and have monitored annual averages greater than 0.3 mg/m3.  The median underestimation factor for these 42 monitors is 16.7.  What caused this underestimation?  We have two theories:

  1. The Re-entrainment Hypothesis: Near sources which have been emitting lead for many years, there may be lead particles in the soil, which can reenter the air when the soil is stirred up by wind or human activities.  There may also be lead in the soil from the days when leaded gasoline was the norm.  These "re-entrained" particles are read by the monitors, but they are not accounted for in the NTI, and thus would not contribute to the model estimates.  We also do not have a background concentration for lead, which could include re-entrained particles.
  2. The Location Uncertainty Hypothesis: This theory was introduced in section V.A. above, but may be especially true for lead.  Many of the lead sources are isolated facilities in rural areas, so the area of high concentration might be very small.  Any errors in emissions locations, release heights, etc., may cause the model to miss the peak, causing an underestimate at the exact monitor location.
The uncertainties involved in the deposition algorithm of the ASPEN model (section IV.B.iii.b) and the uncertainties involved in the estimation of fugitive emissions (section IV.B.ii.a.2) also are likely contributors to underestimation.

We could think of no convenient way to test the first hypothesis.  We can look at the second hypothesis by using MAXTOMON methods and at the percent of defaulted emissions near the source.  We did both.  In order to focus more closely on what caused the underestimation, we looked only at the 30 of the 42 monitors which were underestimated by a factor of 10 or greater.  For the sake of brevity, we will call these the "discordant monitors".

a) MAXTOMON results.  The model receptors used in the MAXTOMON tests included exact monitor locations and tract estimates.  But since many of the lead sources are in rural areas, away from urban areas and their small census tracts, the network of receptors used in the MAXTOMON test is not that dense.  Thus, it is possible that the model could be simulating a peak concentration near the monitor, but none of the tract centroids or monitor locations are near this peak.  For this reason, we feel that the MAXTOMON test is not as effective (without adjustment for receptor network size) in areas of low population density.  Here are the MAXTOMON test results:


Table 10.  MAXTOMON table for the discordant lead monitors.  29 of the 30 discordant monitors have no higher modeled concentrations within 30 km.
Percent Missing Low at Radius Of:
# Discordant Monitors 0 km
(Exact Monitor
Location)
2 km 4 km 6 km 8 km 10 km 20 km 30 km
30 100% 100% 100% 100% 100% 100% 100% 97%

So even a search radius of 30 km brings only one discordant monitor out of "underestimated" status.

b) Location uncertainty results.  The following table shows the percent of emissions falling into each of the four location categories for each of the discordant monitors.  Included are all sources with reported locations within 50 km of the monitor.


Table 11.  The 30 discordant monitors and location categories of nearby point source emissions.
Total Emissions In Each Location Category Within 50 km (1996 tons)
Monitor ID
State
County
Monitored Conc. (mg/m3)
Modeled Conc.
(mg/m3)
Underest. Factor
County Defaulted
Zip Code Defaulted
Not Defaulted, Inside County
Not Defaulted, Outside County
120571066 FL Hillsborough
0.48
0.01
83
1.54
0.00
60.36
0.00
120571067 FL Hillsborough
0.42
0.01
63
1.54
0.00
60.36
0.00
120571071 FL Hillsborough
1.69
0.00
691
0.00
0.00
61.33
0.00
170310068 IL Cook
0.33
0.01
45
0.55
0.00
19.99
0.00
171191012 IL Madison
1.87
0.01
298
0.00
0.00
2.24
0.00
171191013 IL Madison
1.01
0.01
167
0.00
0.00
2.24
0.00
171191015 IL Madison
0.80
0.01
131
0.00
0.00
2.24
0.00
180350008 IN Delaware
0.34
0.03
11
2.16
0.24
6.02
2.74
270370462 MN Dakota
0.46
0.03
17
6.66
0.00
4.80
0.00
270370463 MN Dakota
0.36
0.01
27
6.66
0.00
4.80
0.00
290870006 MS Holt
0.56
0.00
120
0.00
0.00
2.90
0.00
290870008 MS Holt
0.56
0.01
50
0.00
0.00
2.90
0.00
290930016 MS Iron
0.84
0.03
33
2.64
0.00
5.07
0.00
290930020 MS Iron
0.35
0.02
15
2.64
0.00
5.07
0.00
290930021 MS Iron
0.58
0.03
20
2.64
0.00
5.07
0.00
300490714 MT Lewis and Clark
2.21
0.14
16
0.00
0.00
21.52
0.00
300490726 MT Lewis and Clark
1.01
0.06
18
0.00
0.00
21.52
0.00
300490727 MT Lewis and Clark
2.23
0.13
17
0.00
0.00
21.52
0.00
310550049 NE Douglas
3.79
0.10
40
0.01
0.00
21.27
0.00
420110202 PA Berks
0.40
0.03
13
0.06
0.00
5.64
0.00
420110203 PA Berks
0.55
0.03
17
0.06
0.00
5.64
0.00
421010049 PA Philadelphia
0.77
0.06
13
0.18
0.00
1.65
0.01
421010449 PA Philadelphia
3.67
0.04
104
0.18
0.00
1.65
0.01
421010549 PA Philadelphia
0.38
0.04
11
0.18
0.00
1.65
0.01
471570045 TN Shelby
0.99
0.04
24
0.00
0.00
3.83
0.00
471870100 TN Williamson
0.57
0.02
24
0.51
0.00
3.11
0.00
471870104 TN Williamson
0.34
0.02
17
0.51
0.00
3.11
0.00
471871101 TN Williamson
0.62
0.01
56
0.51
0.00
3.11
0.00
480850009 TX Collin
0.42
0.03
13
4.02
0.00
2.26
0.00
550090021 WI Brown
0.48
0.00
494
0.00
0.00
0.40
0.00

Some of the monitors do not have significant emissions nearby at all.  The monitors in Madison County, Illinois; Holt County, Missouri; Philadelphia County, Pennsylvania; and Brown County, Wisconsin have less than 3 tons of lead point source emissions within 50 km.  This could mean that some lead sources are missing in the NTI in these areas.  None of these four states were in the "low participation" category with their point source inventory in section IV.B.ii.a above, however.  Some other discordant monitors have a large percentage of emissions nearby falling into one of the three "uncertain" location categories; for example, the monitor in Indiana has several sources which fall more than 5 km outside the reported county (more on Indiana later), and the three monitors in Minnesota and Texas have a large percentage of county-defaulted emissions nearby.

For some of the other monitors, we were able to do a "before and after" test.  Six of the source-oriented monitors are near three large lead sources which were defaulted in the penultimate model run, but located more accurately in the final model run.  We can compare the agreement between model and monitors before the sources were located accurately to after.  Some other changes to the modeling system were made between the penultimate and final model runs, but we believe the correction of defaulted locations is the most important change affecting the model estimates for these source-oriented lead monitors.

1) The Missouri Source.  The source in Missouri (NTI Site ID ES0912) emits 2.90 tons/year of lead and is close to the two discordant monitors in Holt County, Missouri.  Defaulted, the source was about 10.6 km from the monitors.  The true location of the source is within .7 km of both monitors.  There are no other sources nearby.


Figure 6. Defaulted lead point source in Missouri.
Figure 6.  Defaulted lead point source in Missouri.  The source is located right next to the monitors, but its location was defaulted to 10 km from the monitors.

The following table compares the model-to-monitor agreement before to after.

Table 12.  Model-to-monitor agreement for Holt County lead monitors, before and after source was accurately located.
    Defaulted (Before) Corrected (After)
Monitor ID State County Monitored Conc. (mg/m3) Modeled Conc. (mg/m3) Underest. Factor Modeled Conc. (mg/m3) Underest. Factor
290870006 Missouri Holt
0.56
0.00
311
0.00
120
290870008 Missouri Holt
0.56
0.00
350
0.01
50

The model estimates are still much lower than the monitors, but the agreement is improved.

2) The Tennessee Source.  The source in Tennessee (NTI Site ID ES099) is in Memphis.  It emits 2.70 tons/year of lead and is close to two source-oriented monitors in county 47157 (Shelby County, Tennessee).  One monitor is discordant, i.e., underestimation factor of greater than 10, and the other is not.  Defaulted, the source was about 13 km from the monitors.  Again, the true location is right next to the monitors - within .1 km.  The only other large source in the area (3.2 tons/year) is about 10 km away.

Figure 7. Defaulted lead point source in Tennessee.
Figure 7.  Defaulted lead point source in Tennessee.  The source is located right next to the monitors, but its location was defaulted to 13 km from the monitors.  The correction improved the model-to-monitor agreement considerably.

The following table compares the model-to-monitor agreement before to after for these two monitors.

Table 13.  Before-and-after model-to-monitor agreement for Shelby County lead monitors.
   
Defaulted (Before)
Corrected (After)
Monitor ID State County Monitored Conc. (mg/m3)
Modeled Conc.
(mg/m3)
Underest. Factor
Modeled Conc.
(mg/m3)
Underest. Factor
471570044 Tennessee Shelby
1.80
0.01
327
0.29
6
471570045 Tennessee Shelby
0.99
0.01
181
0.04
24

Agreement here improves quite a bit, even though the source is not that large.

3) The Florida Source.  The source in Florida (NTI Site ID EM3440) is in Tampa.  It emits 0.60 tons/year of lead, and is close to the three discordant monitors in Hillsborough County, Florida.  Defaulted, was about 5.9 km from two of the monitors and 16 km from the third.  Its true location is 4.85 from the pair and 6.12 km from the other.
 

Figure 8. Defaulted lead point source in Florida.
Figure 8.  Defaulted lead point source in Florida.  The correction did not improve model-to-monitor agreement in this case.  It is possible that there is a large point source near the monitors which is missing from or mislocated in the NTI.

The following table compares the model-to-monitor agreement before to after for these three monitors.

Table 14.  Before-and-after model-to-monitor agreement for Hillsborough County lead monitors.
   
Defaulted (Before)
Corrected (After)
Monitor ID State County Monitored Conc. (mg/m3)
Modeled Conc.
(mg/m3)
Underest. Factor
Modeled Conc.
(mg/m3)
Underest. Factor
120571066 Florida Hillsborough
0.48
0.01
72
0.01
83
120571067 Florida Hillsborough
0.42
0.01
56
0.01
63
120571071 Florida Hillsborough
1.69
0.00
565
0.00
691

Because the source is small and its true location is not too far away from its defaulted location, the model estimates did not change drastically, and these monitors are underestimated drastically.  The underestimation factors at all three monitors actually increased slightly.  This is in part due to the other changes in the modeling system between the penultimate and final model runs.  It is possible that a source is missing or mislocated that should appear near these monitors, particularly at monitor 120571071 which has an especially severe underestimate.  There is another source of similar size (0.50 tons/year) which is also defaulted in this county.  We do not know its true location.  There are large sources in the area for which the state reported location had no obvious problems, but these are not reported to be particularly close to the three monitors (about 19 km away).

In the first two before-and-after cases, when we located the source correctly, the model estimates were much improved, but still lower than the monitor averages.  The best agreement we found was an underestimation by a factor of six, for monitor ID 471570044.  This result suggests that location defaulting can contribute significantly to underestimation for source-oriented lead monitors, but it is not the only cause.

One of the most interesting discordant monitors is 170310068.  It is in Chicago, very close to Lake Michigan and the Indiana state line.  This site and the surrounding area are shown in the map below.

There are six large (>1 ton/year) sources in Indiana which fall within 50 km of this monitor.  These are circled in the map.

One possible explanation for the underestimate at this site is that these six large Indiana sources are mislocated.  In general, many of the Indiana sources have reported locations which place them in Kentucky and Ohio, probably south and east of their true locations.  So some of the Indiana sources are offset to the southeast.  If the six large Indiana sources in the circle are moved northwest, they will be much closer to the monitor than they are now, which would presumably increase the model estimate at the monitor site.  Overall, of the 1333 lead emissions sources identified as belonging to Indiana, 155 (11.6%) have latitude/longitude coordinates that place them outside of Indiana, accounting for 5.7% of the emissions.  This would also explain the underestimate at the monitor in Indiana (180350008).

Further support for the theory that these six sources are mislocated comes from the surrounding area.  There are no big cities near the circled sources.  If they are moved northwest, they move into the heavily industrialized cities of Hammond and Gary.


Figure 9. Possible mislocated point sources in Indiana.
Figure 9.  Possible mislocated point sources in Indiana.  Some of the Indiana sources are offset to the southeast.  We wonder if the sources in the circle are offset to the southeast as well.  (Only sources with emissions > 1 ton/year are shown here.)

In light of these results, it is possible that the current state of the NTI’s source location data precludes point-to-point model-to-monitor comparisons for point source oriented monitors, particularly for metals concentrations which tend to decay rapidly with distance from the source due to their high particle size and weight.  We wonder if point-to-point model-to-monitor comparisons for source-oriented monitors should be done only on a small scale, where emissions and source locations can be more accurately characterized.

Lead seems to stand out as the most underestimated pollutant in the comparison.  This may be due to the large number of source oriented monitors, but even the other monitors are underestimated, and the MAXTOMON methods did not significantly change this.  We wonder if there really is reentrainment going on, or something else causing lead emissions to be underestimated nationwide.

ii) Cadmium.

Of the 20 cadmium monitors, 7 are in the Chicago area, and 6 are spread throughout New York.

The 6 New York monitors are all underestimated by a factor of five or greater.  Two of the 7 Chicago monitors are estimated within a factor of 2, with the other 5 being underestimated.  However, there are higher modeled concentrations within 30 km of all but one of these 13 monitors.  Nationwide, 15 of the 20 monitors have higher modeled concentations within 30 km.


Table 15.  MAXTOMON table for cadmium.  For five of the 20 cadmium monitors, no higher modeled concentration can be found within 30 km.
Percent Missing Low at Radius Of:
Pollutant # Monitors 0 km
(Exact Monitor
Location)
2 km 4 km 6 km 8 km 10 km 20 km 30 km
Cadmium 20 85% 85% 75% 75% 75% 60% 35% 25%

iii) Chromium.

Of the 36 chromium monitors, 10 are in Staten Island, NY; 8 are in the Chicago area; and 14 are spread throughout California.

The ten monitors in Staten Island are all underestimated by factors of between 6.64 and 8.30.  The monitor averages are very similar among the ten sites.  There are no census tracts on Staten Island with modeled concentrations similar to the monitors, but there are tracts within 20 km of the monitors with modeled concentrations higher than all 10 monitor averages.

However, the eight monitors near Chicago are generally estimated accurately at the monitor locations.  The model-to-monitor ratios are all between 0.65 and 1.13, so all are easily estimated within a factor of 2, and all but one are estimated within 30%.

The California monitors are estimated inconsistently.  Only 2 of the 14 are estimated within a factor of 2 at the exact monitor locations.  Of the other 12, 2 are overestimated and 10 are underestimated.

These state-to-state differences suggest differences in the state inventories.  The results for cadmium suggest the possibility that New York has low emissions estimates for cadmium as well.  Possible differences in monitoring must also be considered.

VI. Conclusions

At the exact monitor locations, the model estimates were lower than the monitor averages for most of the pollutant/monitor combinations.  Benzene was the only HAP which showed apparent good agreement at the exact monitor locations.  We didn't plan to adjust model estimates based on the results of this model-to-monitor comparison, which we think is reasonable given all the uncertainties involved.  However, for some of the HAPs, especially lead, it seems likely that the modeling system is systematically underestimating monitored concentrations.

By "modeling system", we mean to emphasize that the estimates are the result of:

emission estimates
+
spatial allocation estimates (within each county)
+
dispersion modeling (dispersion, deposition)
+
background estimates.

We tried to use the MAXTOMON test to see whether the model was mislocating the peak concentrations, or systematically underestimating.  On a pollutant-by-pollutant basis, here are the percent of sites underestimated at different radii, out to 50 km:


Table 16.  MAXTOMON table for all pollutants, out to a distance of 50 km.
Percent Missing Low at Radius Of:
Pollutant # Monitors 0 km
(Exact Monitor
Location)
10 km 20 km 30 km 40 km 50 km
Benzene
87
59%
25%
20%
11%*
8%
6%
Perchloroethylene 44 86% 43% 23% 9%*
7%
7%
Formaldehyde 32 88% 56% 31%* 31%
28%
25%
Acetaldehyde 32 91% 56% 38%* 34%
31%
31%
Lead 242 91% 65% 51%* 40%
36%
34%
Cadmium
20
85%
60%
35%*
25%
25%
25%
Chromium
36
83%
39%
28%*
25%
22%
17%

Each pollutant has a certain distance which we starred: in our opinion, this is where the effect of the uncertainty distance starts to wane.  In general, we might say that the modeling system has uncertainty in locating source impacts to within 20 km.  But even as we go out to 50 km, many of the pollutant/monitor combinations are still underestimated for all HAPs except benzene and perchloroethylene.  This suggests systematic underestimation of the aldehydes and metals.

We look forward to receiving the monitoring data collected from the pilot cities.  This monitoring data will have fewer uncertainties surrounding it than most of the data currently in the Archive.  We can compare the monitoring data from this network to future model runs.

In general, we think that the most effective way to improve agreement between the model estimates and monitor averages is to improve the emissions inventory.  One possibility would be to conduct a study on a small sample of sources, to see if the emissions rates are accurately estimated.  But if these model estimates are used on a local scale, we think it is crucial to:

  1. For pollutants dominated by point sources, get better data on the source locations and releases.
  2. For pollutants dominated by area sources, work on either improving our spatial allocation methods or ceasing the pattern of aggregating to the county level in the inventory, then disaggregating for modeling purposes.
  3. Accurately estimate background concentrations, possibly on a regional basis instead of nationwide.
As it stands now, we certainly think it is a mistake to attribute the model estimates to a census tract.  The spatial distribution of concentrations within an area of any size is uncertain.  Any particular model estimate may be too high or too low, compared with actual conditions that existed in 1996.  However, the available data suggest that the model estimate for any particular HAP/location combination is more likely to be lower than actual 1996 conditions, rather than higher.

VII. References

1See https://www.epa.gov/oppecumm/air/air.htm.

2Rosenbaum, A.S., Stiefer, P.S., and Iwamiya, R.K., (1999): Air Toxics Data Archive and AIRS Combined Data Set: Data Base Descriptions.  Prepared for US EPA by Systems Applications International, SYSAPP 98/05r.

3Code of Federal Regulations, Title 40, Part 136, Appendix B, Revision 1.11.

4Curran, T.C., and Steigerwald, B.G., (1982): Data Analysis Consequences of Air Quality Measurement Uncertainty.  Paper presented at the 75th Annual Meeting of the Air Pollution Control Association, New Orleans, Louisiana, June 1982.

5Meade, P.J., and Pasquill, F., (1958): A study of the average distribution of pollution around Stayhorpe.  Int. J. Air Pollution.  Pergamon Press.  Vol. 1, pp. 60-70.

6Lucas, D.H., (1958): The Atmospheric Pollution of Cities. Int. J. Air Pollution.  Pergamon Press.  Vol. 1, pp. 71-86.

7Pooler, F., (1961): A prediction model of mean urban pollution for use with standard wind roses.  Int. J. Air and Water Pollution.  Vol. 4(3/4):199-211.

8Martin, D.O., (1971): An urban diffusion model for estimating long term average values of air quality.  J. of the Air Pollution Control Association.  Vol 21(1):16-19.

9Calder, K.L., (1971): A climatological model for multiple source urban air pollution.  Paper presented at First Meeting of the NATO/CCMS Panel on Modeling.  Paper published in Appendix D of User’s Guide for the Climatological Dispersion Model.  EPA-R4-73-024.  Office of Research and Development, U.S. Environmental Protection Agency, Research Triangle Park, NC 27711, pages 73-105.

10Turner, D.B. and Edmisten, N.G., (1968): St. Louis SO2 dispersion model study - description of basic data.  (Unpublished report, Division of Meteorology, NAPCA).  Atmospheric Sciences Modeling Division, U.S. Environmental Protection Agency, Research Triangle Park, NC 27711.

11Turner, D.B., Zimmerman, J.R., and Busse, A.D., (1971): An evaluation of some climatological models.  Paper presented at Third Meeting of the NATO/CCMS Panel on Modeling.  Paper published in Appendix E of User’s Guide for the Climatological Dispersion Model.  EPA-R4-73-024.  Office of Research and Development, U.S. Environmental Protection Agency, Research Triangle Park, NC 27711, pages 107-131.

12Briggs, G.A., (1969): Plume Rise.  TID-25075. U.S. Atomic Energy Commission, Available from the National Technical Information Service, U.S. Department of commerce, Springfield, VA 22151, 85 pages.

13Holland, J.A., (1953): A meteorological survey of the Oak Ridge area: final report covering the period 1948-1952.  U.S. AEC Report ORO-99.  Technical Information Service, Weather Bureau, Oak Ridge, TN. pages 554-559.

14Irwin, J.S. and Brown, T.M., (1985): A sensitivity analysis of the treatment of area sources by the Climatological Dispersion Model. J. of the Air Pollution Control Association. Vol 35(4):39-364.

15McElroy, J.L. and Pooler, F., (1968): St. Louis Dispersion Study Volume II-Analysis.  National Air Pollution Control Administration.  Publication Number AP-53.  U.S. Department of Health Education and Welfare, Arlington, VA, 51 pages.

16Gifford, F.S., (1976): Turbulent diffusion-typing schemes: a review.  Nuclear Safety. Vol 17(1):68-86.

17Turner, D.B. and Novak, J.H., (1978): User’s Guide for RAM, Volume 1, Algorithm Description and Use.  EPA-600/8-78-016a.  Office of Research and Development, U.S. Environmental Protection Agency, Research Triangle Park, NC 27711, 70 pages.

18U.S. EPA, (1995): User’s Guide for the Industrial Source Complex (ISC3) Dispersion Models, Volume I - User Instructions.  EPA-454/B-95-003a.  Office of Air Quality Planning and Standards, U.S. Environmental Protection Agency, Research Triangle Park, NC 27711, 320 pages.

19Turner, D.B. and Irwin, J.S., (1983): Comparison of sulfur dioxide estimates from the model RAM with St. Louis measurements. Air Pollution Modeling and Its Application II, (Edited by C. De Wispelaere), Plenum Press, pages 695-707.

20Gifford, F.A. and Hanna, S.R., (1973): Modeling urban air pollution.  Atmospheric Environment.  (7):131-136.

21Hanna, S.A., Briggs, G.A., and Hosker, R.P., (1982): Handbook on Atmospheric Diffusion.  Available as DE82002045 (DOE/TIC-11223) from the National Technical Information Service, U.S. Department of Commerce, Springfield, VA, 22161, 108 pages.

22U.S. EPA, (1999): A simplified approach for estimating secondary production of hazardous air pollutants (HAPS) using the OZIPR model.  EPA-454R-99-054.  U.S. Environment Protection Agency, Research Triangle Park, NC 27711, 86 pages.


Local Navigation


Jump to main content.