RE: Processing bulk measurement data
RE: Processing bulk measurement data
- Subject: RE: Processing bulk measurement data
- From: Ethan Hansen <email@hidden>
- Date: Tue, 02 Aug 2016 12:22:59 -0700
- Organization: Dry Creek Photo
Terry,
We perform this type of comparison in two situations. The first is building averaged profiles for media vendors and printer manufacturers. The underlying concern is that a particular printer, media or ink batch, etc. may be an outlier. The second instance sounds more along the lines of what you are looking for. Our largest client has fleets of hundreds of printers from several vendors and different model types. We construct a constantly evolving expected data set for each printer manufacturer/model (or model family) variation. Measurements can then be compared to the expected data and also compared against historical data from the particular machine to detect outlier behavior. The results are fed into both our profiling algorithms and, when significant
Initially we used Excel as a processor. As you found, this quickly becomes impractical. We then switched to home-brewed software that, albeit flexible and powerful, was a PITA to maintain and add features to. We now use SAS JMP. It can easily import custom data sets - if you are already importing CGATS files into Excel, JMP is not going to be any more difficult. The real power, however, are the included statistical tests and graphing routines. Depending on t test values, we typically use either Tukey's HSD or a Thompson Tau test to determine if a measurement is a true outlier at a given confidence level. Cumulative percent plots are also easy to generate; there isn't a faster method to visualize your data, see if there is a non-normal distribution, and check if there is more than one mechanism driving the measured behavior.
JMP isn't cheap (unless you are associated with an academic institution), but we've found it to be worth the price. There is a free (30 days?) evaluation.
Cheers,
Ethan
> From: Terence Wyse
> Looking for recommendations on software that can process fairly large
> amounts of measurement data (10-50+ data sets) with the purpose coming
> up with the optimal data set. Currently doing this in Excel which is
> painful.
>
> Some of the specific features I’m looking for:
> Import or drag/drop a set of measurements, process the AVERAGE of the
> set but also highlight data outliers that should not be included.
> Import or drag/drop a set of measurements, process the MEDIAN of the set
> but also show data outliers that should not be included.
> It should be able to do either of the functions above on a patch-by-
> patch basis rather than including/excluding the entire data set(s).
> Ideally the tool would use a CRF (cumlative relative frequency)-like
> method of displaying the data and the outliers…..and then un-check the
> outliers either manually or based on metrics (dE threshold)…and then
> produce a final data set.
> This isn’t necessarily for the purpose of profiling but more creating an
> ideal set of reference measurements for a fleet of devices.
_______________________________________________
Do not post admin requests to the list. They will be ignored.
Colorsync-users mailing list (email@hidden)
Help/Unsubscribe/Update your Subscription:
This email sent to email@hidden