EOS R1 Autofocus: What Sets It Apart from the EOS R5 Mark II?

This has always been the story of 1 vs 5 series in my experience. The most glaring example were perhaps the 5d3 vs 1dx. All people said was that the AF was the same. But for anyone who used both that juuust wasn’t the case. Same thing goes for comparing and reading just specs. «Everybody» are always underwhelmed when a new Canon body comes out, yet, they perform absolutely stellar and without all the insane caveats of Sony and now Nikon. I don’t know why anyone is surprised the AF in the R1 will beat out both the R3 and R5. It was actually obvious with the specs when it was know in had cross type sensors, it was the big thing being discussed here for years and years, and then almost ignored with all the talking heads said «the R5 II is more of a flagship then the R1» and «I’m not getting an R1 , the R5 II will be better for me» . Using the double cross sensors of the 1dx2 was so much better then older cross type and non cross type.

And just to state again, I feel the R3 was also kind of dismissed and downplayed and ignored by many that went for the R5 and R5 II. A store rep even recommended me the R6 II over the R3 because of the specs saying it’s «basically the same camera». I left. The R3 is the perfect compromise for me and has performed way beyond any other Canon body I’ve owned or tried..
But it's only 24mp
 
Upvote 0
It's wonderful that the R1 has rotated half of the green pixel elements so it can detect both angles (since green has 2 pixel elements in order to fill a simple Bayer square of 4 pixel elements). I wish they had it on the R5m2, but I knew they didn't and I still preferred the R5m2 over the R1 for my particular (non-pro) usage case. I think we can look forward to more of these cross-green sensors in future cameras.

But, I never did truly understand the mechanism (in complete detail!) how the split-pixel idea works to reliably detect focus distance with such a vanishingly small (~2 micron) distance between the center of the pixel element halves. How does it generate a phase difference of the light between each half and how does the front/back focal distance affect it? If anyone knows of a fully detailed article on this then please let me know.
 
Last edited:
Upvote 0
I thought I would weigh in with my experiments and experiences comparing the AF on the Canon 5Ds and the Canon R6. I apologize for the length, and there is a tl;dr at the end.

I'm going to divide AF into two functions: focus tracking which is distance to subject under a selected AF point, and subject tracking which is selecting the AF point over the subject. Some of the 'why' is subjecture on my part, so take it with a grain of salt. An actual camera engineer might come by and tell me I'm wrong on those points.

5Ds AF Design: The 5Ds AF is derived from the 1DX with an improved 150K pixel metering sensor that provides subject recognition (iTR tracking). 61 pts, 41 cross type, 5 double cross type vertical center. There are a couple differences versus the 1DX. In single shot mode the 5Ds will take a little longer to verify and adjust focus since the 50mp sensor will reveal very slight focus 'errors' not visible on an 18mp sensor. Sometimes this includes a secondary lens movement. I do not believe it does this in continuous AF, or at least I have never seen this delay in that mode. If I remember correctly, the 1DX can drive a lens faster on initial acquisation (literally more power to the lens motor), and I do not believe the 5Ds has this ability. There may be other differences such as processing speed or PDAF sensor readout frequency.

R6 AF Design: The R6 has DPAF II which IMHO is the first time Canon mirrorless cameras could compete with upper tier DSLRs on focus tracking. The entire sensor is a massive AF array, and it uses the sensor for both focus and subject tracking. One weakness is that there are no cross type points, which the article mentioned. The camera cannot AF on detail which is perfectly aligned with the AF points and you have to tilt the camera slightly for it to lock.

Focus Tracking: purpose built DSLR AF sensors are very, very good at this, and appear to still be competitive with mirrorless. Using a selected AF point in continuous mode the 5Ds can lock and track subjects as well as the R6 and sometimes even better or faster. A practical example of this: at an airshow I will put the center AF point on a plane when it is still a dot in the distance so that I'm smoothly following it as it approaches flight line. The 5Ds will quickly and confidently lock and hold that dot. With the R6 I found that I had to wait until the plane was a little bit larger in the VF to hit the AF button. If it's too small the R6 may start to hunt, and is slower on initial lock even if it doesn't hunt. (When the plane is larger it's a wash.) In testing, there are definitely small detail situations where the 5Ds can lock faster than the R6, even if the detail is not parallel to the R6 AF points and the R6 does lock. If both have good contrast information for their design then it's a wash except with my Tamron 70-200 f/2.8 G2 where the R6 is faster.

There's an interesting inverse to this situation when detail is very low contrast but the area of detail is larger than a single AF point on the 5Ds. (Imagine a large low contrast pattern, something like single color animal fur.) Then the R6 can lock faster. I would guess that mirrorless cameras are always using a relatively wide array of focus points to try and gather enough contrast information to compute a solution, and then perhaps refine the solution with a smaller set as it gets closer. (I comment on this further below.) So in that scenario it's the R6 that has more and better information to start. This is mitigated somewhat by using assist points on the 5Ds.

I also found that with fast primes and long telephotos (i.e. 85mm f/1.4L IS and 100-400L II IS) the 5Ds is less likely to hunt, which the article mentioned. At a certain point of defocus the 5Ds is still able to drive directly to the subject while the R6 has to start hunting to get closer and then drive straight to it. I have one shot of an F-16 Thunderbird during a surprise afterburner flyby where the 5Ds was completely out of focus when I swung around to capture it, and the AF absolutely nailed it before someone's head got in the way, i.e. I had one chance. I've missed a couple shots like that with the R6, and I've gotten into the habit of manually positioning the 100-400 closer to the correct distance. So if a plane just flew by and another one is coming in the distance, I'll swing the focus ring to infinity before activating AF on the R6.

It should also be noted that while the 5Ds AF is good in low light, the R6 kills it here. The R6 amazes me with its ability to focus on stars, for example. The 5Ds can do it, but it needs a very bright one.

Subject Tracking: the 5Ds is actually pretty good here with iTR on. If you put your initial AF point on a face, then it will use face detect to follow it. If it's not a face, then it will just try to follow the shape/color/brightness that was under the selected point. It can drift if there are similar faces or details nearby, but it's definitely usable and 'mirrorless like.'

The R6 is of course better here. Better subject training, faster processor, and more image data. It can hold an eye and is much less likely to drift. Mirrorless passed up the DSLRs that could to this before it could match them on focus tracking (the R for example). One thing that doesn't get enough mention is that 'None' is a valuable setting. If you want to track something that the camera is not trained for, None will try to hold the shape/color/brightness that was under the initial AF point just like iTR DSLRs, only better. Naturally if the camera is trained for a subject it's more likely to hold onto it, but None works pretty good for everything else.

I have found that when photographing pets up close with a fast prime (85mm f/1.4L), I prefer selecting the point. They can move faster than the R6 can compute a solution given that it has to analyze for subject and then acquire focus. If the R6 isn't busy trying to figure out which point to use, then it can keep up. This is a very challenging situation given the focus throw, shallow DoF, and subject speed, and does not reflect on subject tracking under normal conditions which is excellent. Naturally I always use a selected point here on the 5Ds since there is no animal or eye training.

For the heck of it, I'll throw in the EOS 3. While still fast, it's slower than both the 5Ds and R6 at focus acquistion and tracking. Like the 5Ds it's better at acquiring focus when the lens is well out of focus with the center cross point. While it doesn't have modern image based subject recognition, if the subject is well isolated in depth (i.e. a plane in the sky), it can reliably track it across all 45 points.

What does all this mean? As the article points out, DSLRs use purpose built phase detect AF modules. The PDAF module lens assembly and sensor layout are optimized to give actionable contrast information to the computer. This is more than just cross and double cross points, it's also the size of the line pairs. Mirrorless cameras are limited by the size of the AF points, whether they are DPAF (Canon) or PDAF points embedded in the sensor. With the latter if the AF point is too large you have a hole in your image that can't be interpolated away. It took a long time for mirrorless AF to compete with DSLRs on focus tracking, and I would guess that this is the fundamental reason why. I suspect the computer has to look at many points even in 'single point' mode to get the contrast information it needs, and that the computation is more difficult versus the strong signal from a dedicated AF sensor. And with the exception of the R1, the lack of cross points means it can struggle in some situations, namely aligned detail and extreme defocus.

On the other hand, mirrorless has a lot of image detail to use for subject tracking. They're obviously better than most SLRs and DSLRs which can only subject track if the subject is isolated in terms of depth of field, and also better than the DSLRs which have low resolution image sensors for AE meters and for subject tracking. (Though I would be curious to test a 1DX mark III on this, which would be the last best example of a Canon iTR subject tracking DSLR.)

I'd love to get to handle an R1 and put it through the same tests. Faster readout, faster processor, cross sensors, and I imagine it has the faster initial acquisition of the earlier 1DX models. It's a good bet it would clearly beat the 5Ds in all scenarios except maybe the defocus tests where the 5Ds has the double cross sensors.

tl;dr - recent, upper tier DSLRs are very good at focus tracking and remain competitive with mirrorless thanks to dedicated PDAF sensors with large line pairs and cross/double cross points. Canon iTR DSLRs can subject track, but mirrorless is much better at subject tracking thanks to more image data, better training, and faster processing. R1 may very well eliminate the gap in the few places where upper tier DSLRs still have a small focus tracking advantage thanks to the return to cross points.
 
  • Like
Reactions: 1 users
Upvote 0
I definitely notice my R7 misses focus alot more in portrait orientation than landscape. This may explain that. But sensor readout speed also seems to be important as I've heard alot of ppl comment on a noticeable improvement in AF acquisition and tracking on the R5 vs the R7.
Having said that for the price the R7 is a superb wildlife camera.
Now where's that R5 mk3?
 
Upvote 0
It's wonderful that the R1 has rotated half of the green pixel elements so it can detect both angles (since green has 2 pixel elements in order to fill a simple Bayer square of 4 pixel elements).
I did wonder about that, since Richard stated it but his diagrams showing it appear to be homemade (the 'pixels' are not all evenly spaced, for example). I asked him for confirmation but he didn't reply. Turns out that's exactly the way it is done, even though I haven't seen a patent for that published (Canon has published cross-type AF as alternating orientations of row pairs of DPAF pixels, and as every pixel with QPAF).

But this diagram from Canon Asia is clear. I wonder why Richard didn't use it?

1732580068608.png
 
  • Like
Reactions: 2 users
Upvote 0
I did wonder about that, since Richard stated it but his diagrams showing it appear to be homemade (the 'pixels' are not all evenly spaced, for example). I asked him for confirmation but he didn't reply. Turns out that's exactly the way it is done, even though I haven't seen a patent for that published (Canon has published cross-type AF as alternating orientations of row pairs of DPAF pixels, and as every pixel with QPAF).

But this diagram from Canon Asia is clear. I wonder why Richard didn't use it?

View attachment 221181
This is such a simple (in concept) design to add to the standard Bayer array, that I think it's brilliant that they actually have it in the hands of those with the R1 (like you, I think, Neuro). Enjoy! While the rest of us jabber about what Canon should come out with, you get to strut away down the street with your R1 with 11 million (sorely needed) perpendicular DPAF pixel elements with Dual-Axis-Focus, and the BeeGees singing "Staying InFocus!". :LOL:
 
  • Haha
Reactions: 1 users
Upvote 0
But it's only 24mp
The digital "1" series was always designed for speed and best possible AF, and the target pro photographer group such as sports photographers do not need more resolution. 24 MP even is good enough for glossy print journals (I am editor of a print/online magazine and know what I am talking about).
 
  • Like
Reactions: 2 users
Upvote 0
But, I never did truly understand the mechanism (in complete detail!) how the split-pixel idea works to reliably detect focus distance with such a vanishingly small (~2 micron) distance between the center of the pixel element halves. How does it generate a phase difference of the light between each half and how does the front/back focal distance affect it? If anyone knows of a fully detailed article on this then please let me know.
I don't know either. Is it that each paired split pixel is paired with another split pixel a distance away so that one half of a pixel does the the phase detection with one half of the distant pixel?
 
  • Like
Reactions: 1 user
Upvote 0
I don't know either. Is it that each paired split pixel is paired with another split pixel a distance away so that one half of a pixel does the the phase detection with one half of the distant pixel?
The answer is that it depends. Canon states, "The camera's processor compares the signals from the two photodiodes, and if they match, it knows that this area of the image is in focus. If there is any deviation between them, it looks at pairs of photodiodes across a group of pixels, and can then calculate which direction the lens needs to be adjusted to achieve sharp focus, and how much focus adjustment is required."

So the question is, what is the separation for that 'group of pixels'? The easiest way to think of it is that the greater distance the object is away from the current focal plane (wherever the lens happens to be focused when AF starts), the further separated the split pixels need to be to determine the phase difference. For an object on the focal plane, there is no phase difference. I haven't seen Canon explain this, but this diagram from an academic publication (where the goal was to take what amounts to a DP RAW image and create a full defocus map to generate an all-in-focus image) helps:

DPAF.png
 
  • Like
Reactions: 5 users
Upvote 0
....or...if you learn via cartoons:



A bit on phase detect autofocus.
 
  • Like
Reactions: 2 users
Upvote 0
....or...if you learn via cartoons:



A bit on phase detect autofocus.
The cartoons are a good starter for those struggling with more technical approaches, but they do not give an answer to the question whether the phase difference of only one pixel pair is used by the AF system or also the phase difference between wider separated pixel pairs on the sensor. Like neuro, I guess that Canon uses the latter approach: if the object is far out of focus it gives the AF system two much better separated image parts that need to be overlapped. So it avoids pumping until the system gets a better signal that tells it in which direction it as to drive the AF, what would probably happen in that case when an object is too blurred by de-focus.

In principöe, DPAF works like the rangefinder patch in a classic rangefinder camera or the split image part in a classical SLR focusing screen - but on a micro level.
 
Upvote 0
....or...if you learn via cartoons:



A bit on phase detect autofocus.
I looked at the 2 videos mentioned. The diagram for the path of light in the first was incorrectly drawn as it showed 2 light paths in the same direction and position being bent drastically in 2 different directions by the same micro lens. Since we're not talking about chromatic aberration, this can't be correct. The rest is hand waving and exactly what I'm not looking for. But I do thank you for showing links with us that could be checked out.
 
Upvote 0
The answer is that it depends. Canon states, "The camera's processor compares the signals from the two photodiodes, and if they match, it knows that this area of the image is in focus. If there is any deviation between them, it looks at pairs of photodiodes across a group of pixels, and can then calculate which direction the lens needs to be adjusted to achieve sharp focus, and how much focus adjustment is required."

So the question is, what is the separation for that 'group of pixels'? The easiest way to think of it is that the greater distance the object is away from the current focal plane (wherever the lens happens to be focused when AF starts), the further separated the split pixels need to be to determine the phase difference. For an object on the focal plane, there is no phase difference. I haven't seen Canon explain this, but this diagram from an academic publication (where the goal was to take what amounts to a DP RAW image and create a full defocus map to generate an all-in-focus image) helps:

View attachment 221199
Thanks for this image. It does give a simple idea of what they're trying to do. But the programmer in me wants to see a lot more detail regarding how things are built and real world test examples. One question this brings to me is whether each pixel "bin" is separated from the others by a sort of wall, and if so adding this level of lensing would cause much of the light in a real image to be blocked instead of reaching the sensor area and thus I would expect a large loss of light being sensed. If there is no wall blocking, then there would be a lot of smearing of light between neighboring pixel elements, and I'd like to see some actual graphs of how a single point source of light (only) in sensed by a line of dual pixels. So I would expect to see some loss of light being sensed as well as some loss of contrast and resolution.

But this image doesn't show how "phase" is affected, as it shows how the direction of light on left/right of the main camera lens is sensed by an array. Phase sensing is where 2 versions (one delayed in time/travel more than the other) of the same signal are combined and sensed at the same point so that one maximum interferes with the other minimum and thus the signal is cancelled, and it is the amount of cancellation that is sensed. Maybe "phase" isn't really what is being detected and another word is more appropriate?

Of course, we know that it does work, so that's not the issue. I'm just interested in the details which go beyond simple images like this. But it is a good start and is similar to what I've seen in the past which does say how a line of pixels have to work together to get a result.
 
Last edited:
Upvote 0
I went out and shot my son working out / running at the track this morning and the AF tracking is simply incredible. It would lock on to his head from one end of the track to the other, about 350-400 feet away, I do not recall the R3 being able to do that from that distance. One more thing to note: the diopter adjustment is a push to lock WHICH I LOVE! I cannot tell you how many times I've accidentally turned that thing while getting it out of my case/backpack only to put it up to my eye and have everything blurry.
 
Upvote 0
The cartoons are a good starter for those struggling with more technical approaches, but they do not give an answer to the question whether the phase difference of only one pixel pair is used by the AF system or also the phase difference between wider separated pixel pairs on the sensor. Like neuro, I guess that Canon uses the latter approach: if the object is far out of focus it gives the AF system two much better separated image parts that need to be overlapped. So it avoids pumping until the system gets a better signal that tells it in which direction it as to drive the AF, what would probably happen in that case when an object is too blurred by de-focus.

In principöe, DPAF works like the rangefinder patch in a classic rangefinder camera or the split image part in a classical SLR focusing screen - but on a micro level.
It’s not a guess, it's optics. In-focus subjects have no phase difference, out-of-focus subjects do, and that difference is proportional to the magnitude of the defocus. The phase difference is a vector quantity – it has magnitude and direction, so the AF system can calculate how far and in which direction to move the focal plane. That’s the fundamental principle of phase detect AF.

In a DSLR, the dedicated AF sensor has lines of pixels to accommodate the spread of the phase difference. In a DPAF sensor, separated pixels are used for the same purpose. The separation is not that large a distance for PDAF (a few pixels apart), in a DSLR AF system there is a secondary image-forming lens in front of the AF sensor that is some distance away, which means greater physical separation of the phases.
 
  • Like
Reactions: 2 users
Upvote 0