The second and much more difficult problem (at least if you had no one to turn to for help at the time, which I didn't) was to generate a maximally spaced sampling grid of 1000 points spread out over the entire spheroid surface. I began with a guess-timate assignment of 1000 points at particular whole longitude/latitude combinations, with successive manual adjustments based on the mean arc distances (and variances) that could be calculated from each point to all the others. After just a few iterations those general parts of the spheroid which clearly had "too many" sample points quickly evaporated, leaving me with perhaps another dozen iterations to get things down to a pretty close representation of maximally-spaced sampling. I do not mean to suggest that these exact approaches should be applied in a possible next round of more precise analyses, but I am quite confident that these starting points were more than adequate for present purposes, as (1) I only wanted to get a general idea of what would play out, and (2) the data I would connect to them in the real world analyses are in any case comparatively less precise than this anyway. Only one kind of simulation was attempted here. I first randomly assigned class membership in one of four groups to each of the 1000 points, then calculated the spatial autocorrelation properties of the resulting sets. The resulting four by four matrices of values were then double-standardized. I employed three different variations of the spatial autocorrelation algorithm, but with this many points involved, random assignment is a very powerful influence, and very little organization emerged in the relation of the groups to one another. This was easily seen, as all the input matrices calculated consisted, regardless of the spatial autocorrelation algorithm, of values that varied from one another by only a few percent. A typical input matrix was:
I ultimately ran a total of 1500 double-standardizations (500 for each spatial autocorrelation measure); 16, 16, and 19 of the four by four input matrices eventually passed the test of producing the right type of symmetric double-standardized values. This comes to about 3.4 percent. Thus, rather more pass the test on this kind of surface than under conditions in which purely random numbers are fed into the d.-s. algorithm, and probably around the same as on arbitrarily delimited flat surfaces. There are several important things to take away from this, but none of them involve absolute verdicts. First, and most importantly, even when faced with near total randomness of pattern, some of the resulting pattern summaries (i.e., 4 by 4 matrices) calculate out as the kind of structure that projects as a three dimensional space. Second,
however, the exact number that Further, it is very likely that other class assignment algorithms will produce very different results altogether. I only used a random form of assignment here, but one can imagine any variety of other types (and, of course, on variously shaped surfaces). One such exercise follows in the next section. _________________________
Copyright 2006-2014 by Charles H. Smith.
All rights reserved. Feedback: charles.smith@wku.edu |