E-M1X vs G9 IBIS at 12mm

whumber

Senior Member
Messages
4,617
Solutions
6
Reaction score
6,297
Location
Rutland, VT, US
The claim in the E-M1X review that the G9 IBIS is superior to both the E-M1ii and E-M1X was something I just couldn't make sense of based on my experience owning all three of those cameras. I've been trying to get details of how DPReview is conducting their test, but outside of the extremely vague description they include in some of the reviews they seem to have no interest in responding to any of my requests for more details. I decided to try actually measuring the difference between the two cameras to see if my overall impressions matched up with an attempt at objective measurements.

To take some measurements, I'm using a standard ISO-12233:2000 test chart, or at least the central portion of it, to do some slant edge MTF calculations. I have the central portion of the chart printed at 13"x19" which would correspond to a full test chart around 33" wide. I'm using a variable LED light stand so that I can maintain a constant aperture throughout all the tests.

All shots are taken while standing, unsupported, 6.5 ft from the test chart with the camera at the same level as the center of the chart. Best focus is attained by using each cameras "punch in" feature and focusing until a best focus is attained and then a minimum of 50 shots are taken in burst mode. Electronic shutter is used in all cases for best sharpness except on the G9 for the 2s exposures as electronic shutter is not available, in that case the mechanical shutter with electronic first curtain is used. All images are taken with the Olympus 12-40mm f/2.8 lens at 12mm | f/5.6| ISO 100; I chose f/5.6 to attempt to minimize the effect of forward and back sway while not going too far into diffraction territory.

To prepare for processing the results I also took a series of tripod mounted images in order to establish control values as I expected to see significant differences in the MTF calculations between the two cameras just due to factors such as the AA filter and any hidden processing that might occur from the Adobe RAW converter. Fortunately, the two ended up showing almost identical MTF values, although all the MTF values are normalized by the maximum MTF values measured from the tripod control images. For the MTF calculations I used an older version of Reikan Focal's "Focus Consistency" tool where it allows you to load files manually (although the EXIF has to be hacked for my version to make it think that you're using an older Canon or Nikon body). If I was going to be doing this on a larger scale I'd put something together in matlab but this was easy for something quick and dirty since the control tests showed credible results.

Tests were done at 1/10s, 1/5s, 1/2.5s, 1s, and 2s. I only performed the tests with IS enabled as I'm only concerned about the relative performance of the E-M1X and G9 rather than the absolute degree of stabilization relative to an unstabilized system. The results are summarized in the chart below. One important note, the green line represents the MTF value above which I would consider very sharp while the red line is the MTF value threshold I would consider the image to be soft but useable; anything below the red line is blurred. These values are something I had to set arbitrarily but I believe I was fairly aggressive in setting the limits. To give an idea of what is considered sharp, soft, and blurred I'm including images that correlate to each of the thresholds.

Sharp
Sharp

Soft
Soft

Blurred
Blurred

2fc0249bb11d4845a984f43ee1366499.jpg.png

The main takeaway is that, just like I suspected, the IBIS in the E-M1X is superior to the G9. In addition, DPReview's reporting showed that at 0.2s they saw approximately 20% sharp images, 30% soft images, and 50% blurry. In my testing, and this is with just the IBIS rather than IBIS + OIS of the 12-100, I see 0 out of 54 images fall below the sharp threshold. At 0.2s we also see the E-M1X and G9 performing very similarly with a slight advantage to the E-M1X on average. Note for the histograms below, I had to artificially add a count in the lowest and highest bin for each plot simply for reasons of laziness when creating the plots. Also the green, yellow, and red parts of the plots represent sharp, soft, and blurred zones respectively with the red zone truncated to a single histogram bin.

b7342fbcfc23416f954f567ca63a17f4.jpg.png

Similarly at 0.5s they report 0 sharp images, 20% soft, and then the rest blurred. I see 60 above the sharp threshold, 2 soft, and 0 blurred images at 0.4s. Why 0.4s instead of 0.5s? Because I did the test at 0.4s and didn't feel like going back and reprocessing everything, =). At 0.4s we also see the E-M1X really start to pull ahead of the G9. The E-M1X maintains almost a perfect sharp rate but the G9 is down to 64% sharp.

3f0bc8b556fa4bfc8cbd309046187eb1.jpg.png

At 1s, the E-M1X is finally starting to show some weakness as it gets its first truly blurred images but it still maintains a 71% sharp rate. The G9 on the other hand really starts to struggle at this point. It still manages to get sharp images but the sharp rate is down to 27% and the blurred rate is 57%.

8a1d19e7aaea4db5bc69beb0664eec21.jpg.png

Finally at 2s the G9 seems to have given up. I didn't get a single image out of 50 that classified as sharp or even soft, all were well into the blurred range. The E-M1X on the other hand is definitely struggling at 2s but is still managing a 33% sharp rate. Interestingly, there's a bimodal behavior in this range; it seems like the images either come out sharp or completely blurred with very little in between.

9891c685817a497a9aaf2d502b8c10b6.jpg.png

Now before the first person comes out and starts screaming that they can get 2s exposures on their G9 while dancing the jitterbug I'm not saying it's not possible, just that with IBIS only at 12mm and a 6.5ft focus distance it's much easier to do with the E-M1X. After about a year with the G9 I think I've got a pretty good feel for it but if someone knows a secret setting that will greatly improve the IBIS performance I'd certainly be interested.

This testing methodology is certainly not perfect, but I honestly don't understand what DPReview is doing with their test to get the results they publish. My best guess is that they're doing something goofy like standing right next to the test chart when doing the 24mm equivalent measurements which will change translational camera movements from a secondary driver of blur to a primary driver. I've asked them repeatedly for details about their test procedure but I'm always left hanging. Maybe they'll open up their methodology in the future, but in the meanwhile I believe there's some fundamental methodology flaw in their testing and I would recommend taking anything they report with a huge grain of salt.

EDIT: DPReview is doing something silly with most of pictures so you'll have to click on "View Original Size" below the picture to actually see anything.
 
Last edited:
Awesome, thanks for the detailed testing.

Were you able to test the E-M1X against the E-M1.2? Do you think the IBIS is much improved?
I did a few tests with the E-M1ii but didn't publish those results. There does seem to be an improvement but it only shows up in extremely long exposures (i.e. >= 1s). Below that they seem very similar.
 
is the one you can't test for, the person holding the camera. Some people may find holding the G9 stable is easier than the E-M1, and vice versa. For me any differences are so minimal as to be not worth worrying out, lets just say they both have the best IBIS out there bar none, and be happy we have a choice between two excellent cameras.
 
Excellent testing. I expect the EM1X would still have a significant advantage but it would be interesting to see how much closer the G9 might come with a dual-IS enabled lens like the 12-35.
 
Nice work.
This testing methodology is certainly not perfect, but I honestly don't understand what DPReview is doing with their test to get the results they publish. My best guess is that they're doing something goofy like standing right next to the test chart when doing the 24mm equivalent measurements which will change translational camera movements from a secondary driver of blur to a primary driver.
I think this is important. I first discovered it when I was testing my 12-40: IBIS was inexplicably better at 40mm than 12mm. It took me a while to work out why. However, presumably the G9 is subject to the same challenge as the EM1x.
I've asked them repeatedly for details about their test procedure but I'm always left hanging. Maybe they'll open up their methodology in the future, but in the meanwhile I believe there's some fundamental methodology flaw in their testing and I would recommend taking anything they report with a huge grain of salt.
I agree. I notice two things:

- Their sample size of 10 shots (presumably per SS/FL combination) is rather small. The variance of a categorical measurement (sharp/unsharp) is surprisingly large. That's partly because you're throwing out a lot of the information available (sharpness that you could measure on a continuous scale, MTF) by reducing it to a subjective judgment.

- Worse, I think they have a systematic error problem: if you don't use the same shooter and the same rater, the categorization of performance is likely to differ from day to day. Even if you use the same person, their performance will drift, even in a single experiment due to fatigue and learning. The only saving grace is that they simultaneously run a no-IS test as a control. Even with that, the order of the tests will matter.

It would be interesting to look at the variation in no-IS tests across several reviews. I bet it's large. How else would you explain that the EM1x is .7 stops worse than the EM1ii at 200mm?

Sadly, I think the bottom line is that the DPR test is near-useless for cross-review comparisons. They should own up to that and improve their methods in the future.
 
Nice work.
This testing methodology is certainly not perfect, but I honestly don't understand what DPReview is doing with their test to get the results they publish. My best guess is that they're doing something goofy like standing right next to the test chart when doing the 24mm equivalent measurements which will change translational camera movements from a secondary driver of blur to a primary driver.
I think this is important. I first discovered it when I was testing my 12-40: IBIS was inexplicably better at 40mm than 12mm. It took me a while to work out why. However, presumably the G9 is subject to the same challenge as the EM1x.
I've asked them repeatedly for details about their test procedure but I'm always left hanging. Maybe they'll open up their methodology in the future, but in the meanwhile I believe there's some fundamental methodology flaw in their testing and I would recommend taking anything they report with a huge grain of salt.
I agree. I notice two things:

- Their sample size of 10 shots (presumably per SS/FL combination) is rather small. The variance of a categorical measurement (sharp/unsharp) is surprisingly large. That's partly because you're throwing out a lot of the information available (sharpness that you could measure on a continuous scale, MTF) by reducing it to a subjective judgment.

- Worse, I think they have a systematic error problem: if you don't use the same shooter and the same rater, the categorization of performance is likely to differ from day to day. Even if you use the same person, their performance will drift, even in a single experiment due to fatigue and learning. The only saving grace is that they simultaneously run a no-IS test as a control. Even with that, the order of the tests will matter.

It would be interesting to look at the variation in no-IS tests across several reviews. I bet it's large. How else would you explain that the EM1x is .7 stops worse than the EM1ii at 200mm?

Sadly, I think the bottom line is that the DPR test is near-useless for cross-review comparisons. They should own up to that and improve their methods in the future.
What is sad is there faulty reviews influence what people buy
 
Thanks for taking the time to do this.
 
It would be interesting to look at the variation in no-IS tests across several reviews. I bet it's large. How else would you explain that the EM1x is .7 stops worse than the EM1ii at 200mm?
An example: take a look at the Z7/A7iii stabilization shootout:

https://www.dpreview.com/articles/5876118090/image-stablization-showdown-nikon-z7-vs-sony-a7r-iii

There are large differences in unstabilized performance, even though these were shot side by side. You might expect slight differences due to the mass of the system, grip shape, etc., but not like this. Also, their verbal assessment appears to rely on the point at which the camera maintains 50% usable shots, which is an extremely noisy metric given their methods.
 
Last edited:
Here's the thing. You did this test because of your bias towards a particular model, with the intention of proving said bias. As such, the credibility of the results suffers, even if you had every intention of being impartial while performing your test.

Furthermore, not performing unstabilized tests was clearly a mistake because the effectiveness of ibis is measured in comparison to unstabilized results on the same camera body. The Em1X being significantly heavier should lead to better baseline performance typically which would explain this discrepancy.

Anyway, this is rather pointless. It's well known that both cameras have the best stabilization in the market and I really doubt it affected many buying decisions between them.
 
Here's the thing. You did this test because of your bias towards a particular model, with the intention of proving said bias. As such, the credibility of the results suffers, even if you had every intention of being impartial while performing your test.
My impression is that he did the test because there's a strong smell of BS in DPR's methods, irrespective of the results. Since likes both cameras enough to own them, bias is merely your supposition - it could run either way.

Unstabilized tests are useful, but since the EM1x gets knocked for its weight, shouldn't it also get some credit where that weight is useful?
 
Last edited:
Here's the thing. You did this test because of your bias towards a particular model, with the intention of proving said bias. As such, the credibility of the results suffers, even if you had every intention of being impartial while performing your test.

Furthermore, not performing unstabilized tests was clearly a mistake because the effectiveness of ibis is measured in comparison to unstabilized results on the same camera body. The Em1X being significantly heavier should lead to better baseline performance typically which would explain this discrepancy.

Anyway, this is rather pointless. It's well known that both cameras have the best stabilization in the market and I really doubt it affected many buying decisions between them.
Why is it pointless to compare mFTs cameras on a forum for mFTs cameras? While differences in stabilization may not be the most important difference between cameras, it is almost certainly one of the factors for some.

Everyone has bias and there are almost certainly differences in individuals which may result in some types of stabilization being better than others.

If you want unbiased, then CIPA (while not a human user measure) is unbiased. The OP's results appear to agree with the CIPA results where the E-M1X measured 7 stops with a lens without stabilization and 7.5 stops with dual stabilization vs the G9 with 6.5 stops with dual stabilization.
 
Here's the thing. You did this test because of your bias towards a particular model, with the intention of proving said bias. As such, the credibility of the results suffers, even if you had every intention of being impartial while performing your test.
My impression is that he did the test because there's a strong smell of BS in DPR's methods, irrespective of the results. Since likes both cameras enough to own them, bias is merely your supposition - it could run either way.
The OP says right in the first paragraph that this test was performed because he disagreed with the results of the review and wanted his own measurements. How is bias my supposition?
Unstabilized tests are useful, but since the EM1x gets knocked for its weight, shouldn't it also get some credit where that weight is useful?
Heavier cameras are more stable. This does not mean their IBIS performs better. If you put it on a monopod it'll perform even better but that's not the point, is it?

Don't get me wrong, I fully expect the EM1X to give better results than the G9 or any other camera for that matter, exactly because its heavier. But that's not due to a more advanced IBIS system, rather than good old gravity.
 
Here's the thing. You did this test because of your bias towards a particular model, with the intention of proving said bias. As such, the credibility of the results suffers, even if you had every intention of being impartial while performing your test.
Bias is in you!

He did try to get DPR reveal their test method and because they didn't, he made own that is more presentive of the real situation (6.5 feet with 12mm) and did so for both
Furthermore, not performing unstabilized tests was clearly a mistake because the effectiveness of ibis is measured in comparison to unstabilized results on the same camera body. The Em1X being significantly heavier should lead to better baseline performance typically which would explain this discrepancy.
No one cares about unstabilization as the end result shows that Olympus is far better than Panasonic with the test parameters and the size or weight doesn't matter as what matters are the shutter speed, focal length and the results of the "Sharp" or "Soft".
Anyway, this is rather pointless. It's well known that both cameras have the best stabilization in the market and I really doubt it affected many buying decisions between them.
Yes, you are pointless. It is well known that Olympus still has the best stabilization system in the industry, and Panasonic doesn't have better nor be par with Olympus.
 
Why is it pointless to compare mFTs cameras on a forum for mFTs cameras? While differences in stabilization may not be the most important difference between cameras, it is almost certainly one of the factors for some.

Everyone has bias and there are almost certainly differences in individuals which may result in some types of stabilization being better than others.

If you want unbiased, then CIPA (while not a human user measure) is unbiased. The OP's results appear to agree with the CIPA results where the E-M1X measured 7 stops with a lens without stabilization and 7.5 stops with dual stabilization vs the G9 with 6.5 stops with dual stabilization.
Whether the G9 is slightly better as dpreview claims or the EM1X is slightly better as the OP claims, they're very close and it's not a point worth emphasizing imho.

So instead of obsessing about it and asking the OP to re-test with unstabilized control samples as a result of my remarks, I concluded that I don't think it matters.

They're not even in the same price bracket. I don't believe anyone would ever consider a $1.8k price hop, which is more than double the price, towards either model because of a potential quarter stop IBIS effectiveness on this side of the scale or the other.

But hey, don't let my common sense stop you if you feel otherwise.
 
Why is it pointless to compare mFTs cameras on a forum for mFTs cameras? While differences in stabilization may not be the most important difference between cameras, it is almost certainly one of the factors for some.

Everyone has bias and there are almost certainly differences in individuals which may result in some types of stabilization being better than others.

If you want unbiased, then CIPA (while not a human user measure) is unbiased. The OP's results appear to agree with the CIPA results where the E-M1X measured 7 stops with a lens without stabilization and 7.5 stops with dual stabilization vs the G9 with 6.5 stops with dual stabilization.
Whether the G9 is slightly better as dpreview claims or the EM1X is slightly better as the OP claims, they're very close and it's not a point worth emphasizing imho.
Don't ignore the only standardized, unbiased repeatable measure (CIPA). The difference may be important, depending on your lens choice. It would not be most important difference for me, but I appreciate that the OP was willing to do comparison and I don't pretend to know how important it is to others.
So instead of obsessing about it and asking the OP to re-test with unstabilized control samples as a result of my remarks, I concluded that I don't think it matters.

They're not even in the same price bracket. I don't believe anyone would ever consider a $1.8k price hop, which is more than double the price, towards either model because of a potential quarter stop IBIS effectiveness on this side of the scale or the other.
I agree, that stabilization alone would not be the most important reason that I might consider the E-M1X instead of the G9, but the effectiveness of the dual stabilization with the 12-100 f4 would be a factor.

For me, the three Electronic AF Limiter settings on the E-M1.2 (and E-M1X) are the most important difference in the Olympus and the G9 and one of the reasons I would choose the Olympus over the G9. For others, the choice between cameras is probably more affected by differences in video, size, ergonomics, etc..
But hey, don't let my common sense stop you if you feel otherwise.
 
Here's the thing. You did this test because of your bias towards a particular model, with the intention of proving said bias. As such, the credibility of the results suffers, even if you had every intention of being impartial while performing your test.
I don't disagree that I have some bias, or at least an expected result. I don't, however, have a preferred outcome. If it turned out that the G9 was in fact better for longer exposures, as DPReview suggests, then it's something I would very much like to know since it helps me decide which camera to bring with me in different situations. That was the whole point of the test though, to see if my expectation had a basis in reality. I suppose I could have had an espresso before doing all the G9 tests but then that would kind of defeat the point of doing the test.
CIPA, post: 62660831, member: 1644685"]
Accordingly, theoretically speaking, the shake amount when the image stabilization function is OFF should not be any different among cameras except for subtle individual differences. In fact, it has been confirmed through the results of measurements performed on many cameras for this standard that there is no significant difference among cameras.
Anyway, this is rather pointless. It's well known that both cameras have the best stabilization in the market and I really doubt it affected many buying decisions between them.
I personally don't care how it affects the buying decisions of others. I just saw what seemed like a bizarre result from DPReview and wanted to see if I could recreate it.
[/QUOTE]
 
Last edited:
Here's the thing. You did this test because of your bias towards a particular model, with the intention of proving said bias. As such, the credibility of the results suffers, even if you had every intention of being impartial while performing your test.
My impression is that he did the test because there's a strong smell of BS in DPR's methods, irrespective of the results. Since likes both cameras enough to own them, bias is merely your supposition - it could run either way.
The OP says right in the first paragraph that this test was performed because he disagreed with the results of the review and wanted his own measurements. How is bias my supposition?
Thinking DPR's review is wrong on the basis of experience with both cameras, as the OP said, is not the same thing as having a bias for a particular model, as you said.

Anyway, lets suppose that you are right, and posit that the OP is an EM1x fanbois and hates his G9. That (or less) opens up the possibility of bias, but we don't have any particular theory about what went wrong in his test. By contrast, there are several obvious flaws with DPR's methods. So what would you say about the relative credibility of the two?
Unstabilized tests are useful, but since the EM1x gets knocked for its weight, shouldn't it also get some credit where that weight is useful?
Heavier cameras are more stable. This does not mean their IBIS performs better. If you put it on a monopod it'll perform even better but that's not the point, is it?
Actually, it is the point. I don't care how the camera performs relative to its unstabilized self. If the weight and the grip gave me an extra half stop, I'd consider that part of the package. (Unfortunately, the intrinsic penalty of the weight is not one I'm willing to bear.)
 
Last edited:
Seems the DUAL IS 2 would work differently with the Panasonic GX vario version II lenses.
 

Keyboard shortcuts

Back
Top