NOAA/NWS/NCEP/Environmental Modeling Center has a request for data out, one which gives anyone near water who can read a thermometer a chance to do some science. There's a science history behind why this is a request and I'll give my own, biased, view of that.
All data have errors and are messy. Though George Box's comment is often repeated at modelers ("All models are wrong, some models are useful.") it is equally applicable to data and data analysis. All data are wrong, some data are useful.
In the case of sea surface temperatures (sst), efforts to analyze global ocean sst started with badly distributed data sources -- ships. They give you a fair idea of what the temperature is along the paths the ships take. So the ship route between New York and London is pretty well-observed by ship and has been for a long time. But not many ships go through the south Pacific towards Antarctica. If you want to know what's happening down there, you need a different data source. One such is buoys. Though, again, buoys are distributed in a biased way, being mostly towards shore (so that they can be maintained and repaired).
Then came satellites and all was good, eventually, for a while. Polar orbiting satellites see the entire globe. Starting with instruments launched in the early 1980s, it has been possible to make pretty good analyses of global sst, at least on grid cells 50-200 km on a side. Since that is as good or better than any of the ship+buoy analyses could do, that was a great triumph. The ship and buoy data, though, remained and remains important. One of the problems that satellite information faces is that the instruments can 'drift', that is, read progressively too warm, or too cold. To counter that possibility and other issues, the surface data (in situ data) is used as a reference. So for a time in, say, the early 2000s, all was good.
But both scientists and users of scientific information are never satisfied for long. For sst, some of the users are fishermen -- some fish have very particular temperature preferences. As it became possible to do a pretty good global 50 km analysis, with new data over about 2/3rds of the ocean every day, scientists and users started demanding more frequent updates of information, and on a finer grid. They also got increasingly annoyed about the parts of the ocean that only got new observations every 5-20 days. This includes areas like the Gulf Stream, where it is often cloudy for extended periods. The traditional satellites are great, but don't see through cloud.
Another major user of sst information is numerical weather prediction. When weather models were using cells 80-200 km on a side, the sst at 100 km (say) was a pretty good match. But weather models continued to push to higher resolution, so that by the early 2000s, 10 km grids weren't unheard of. The reason for such small grid spacing in weather prediction models is that weather 'cares' about events at very small scale. If weather cares about those smaller scales, then it become important to provide information about sst at the smaller scales. An inadvertent proof of that was made when a model made a bad forecast for a December 2000 storm, and the cause was traced back to an sst analysis that was too coarse. See Thiebaux and others, 2003 for the full analysis.
Plus, of course, there is interesting oceanography that requires much finer scale observations than 100 km. So a couple of different efforts developed. One was to use microwave data to derive sea surface temperatures. AMSR-E was the first microwave instrument used for sst in operations, as far as I know. (Sea ice isn't the only thing that you can see with microwaves!) That addressed the issue of seeing the Gulf Stream (and other cloudy areas) most days. The other was to start pushing for higher resolution sst analysis. This lead to an international effort to analyze the global ocean at high (say 25 km and finer, sometimes 10 km and finer) grid spacing. More is involved in that than just changing a parameter in the program. (You'll get an answer if you do that, but it won't be as good as you had at the coarser grid spacing).
On the ocean side, this quality of the high resolution analyses is doing relatively well. But as you go to finer grid spacings, new matters appear. The Great Lakes are very large, so they can be seen by satellite easily, and they have buoy data through at least part of the year, so that the satellite observations can be corrected at need. But ... go to a finer grid spacing weather model and you discover that there are a lot of lakes smaller than the Great Lakes. For a 4 km model, there are some thousands of lakes just in North America. And none of them have buoys, and almost none even have climatologies. Also at this grid spacing, you start seeing the wider parts of rivers.
Here's where an opportunity arises for people who live near a shore (whether river, lake, or ocean). NOAA/NWS/NCEP/Environmental Modeling Center is requesting observations of water surface temperatures to use as a check on their analysis of temperatures in areas close to shore. (close equals, say, up to 50 km (30 miles), and at least 400m (quarter mile) from shore). Check out the project's web page at Near Shore Lake Project
As always, I don't speak for my employer or any groups I might be a member of. I'm pretty certain that all people who work on sst would disagree with at least parts of my above mini-history. Be that as it may, it should be a fun project.
Delar av Stora Barriärrevet har dött
7 hours ago