Here are my 2 maps1 for this discussion. I used linear regression to predict corn yield for each county in the US, using time (year) as the independent variable. I used the years 1950 to 1969 to create the model, and to predict corn yield in 1970. This should be a reasonable estimate of the ‘expected yield’ for 1970 for each county, if it had been a ‘normal year’.
I then computed the difference between the expected yield and the yield obtained by farmers, and expressed that as the percentage of the expected yield. Negative numbers mean that yields were lower than expected in a county, positive numbers mean that they were higher than expected. Counties with data for less than 9 years were excluded.
1970 corn yields were indeed much lower than expected in the southeast. Corn blight hit very hard. But also note that yield was stable or up in the north and in the west, and look were US corn was grown in 1970. The map below expresses corn area as the percentage of the total area of a county.
Most corn is grown in the corn-belt. The southern parts of it were much affected by the disease (The Illinois Secretary of Agriculture’s estimate that, by August, 25 percent of his state’s corn crop had been lost to the blight may have been spot on). But 1970 was a normal or good year for corn yield in the northern and western parts of the corn belt, and that compensated for the losses incurred elsewhere. If you sum it all up, corn production was about 15% lower than what could have been expected. That is whole lot of corn — but perhaps not that exceptional as far as bad years go.
Here is a table of estimated corn yield by state, as percentage of the expected yield for 1970, and the corn area, as percentage of the national area (only for states with more than 1% of the national corn area in the counties data set).
- Quick & dirty, without cross checking the numbers, but I think the maps speak for themselves. [↩]