Benefits of providing both solvent and focus masks

I’ve long been puzzled by the pros and cons of providing both solvent and focus masks for 3D classification, compared with using only one mask covering the ROI as is implemented in Relion. I can understand the former is a cheap alternative to particle subtraction, in the sense that it accounts for the signal and CTF of the region outside the ROI that bleed into it. In my experience, providing both masks barely separates the particles. The particle distribution is virtually uniform and all the class averages look highly similar. In contrast, if I provide just the solvent mask covering the ROI, the particles are much more efficiently separated.

To put it in more context: I have 4 million particles. Each particle is a segment on a filament. The solvent mask is a soft mask covering the 5 subunits in the center (43 kD for each subunit, 215 kD in total). The focus mask covers the ROI, where we anticipate to see the binding of a 7 kD domain. The solvent mask is generous so that it completely covers the focus mask. About 5% particles are bound by the 7 kD domain and they are the ones I’m hoping to find by 3D classification. Below is how the masks look like:

Below is the final particle distribution with both masks provided:

However, if providing just the solvent mask covering the ROI, that is using the focus mask from the previous job as the solvent mask, the particles are much better separated:

Right now running 3D classification with just solvent mask seems working fine. But I’d like to understand why using both masks performs worse, as it’s counterintuitive.

On a related note, how is the per-particle scale computed if just a small solvent mask covering the ROI is used? I’ve noticed setting the parameter to none consistently performs better than optimal. The documentation says setting it to “none” is only recommended for very small particles or strange cases. I was wondering what are those “strange cases”? Could it be more specific? And in the case of using only the solvent mask, does “very small particle” actually refer to very small mask?

3 Likes

Very interesting observations.

With regards to scale factors, I would usually use the “input” mode, assuming that scale factors have been refined in one of the upstream refinement jobs.

1 Like

Hi @olibclarke ,

Thanks for your suggestion! I have never enabled “minimize over per-particle scale” in any of my refinement jobs, so the input scale factor would still be 1. I will run another refinement with that option on and see if it makes any improvements in 3D classification.

The reason I was reluctant to refine per-particle scale and use the input values is because cryosparc guide noted that it may reduce the job’s ability to separate classes with missing/additional density, which is exactly what I want. On the bright side, a 7kD domain probably won’t make much change to the intensity/contrast of the particle, so it’s definitely worth trying. Another reason is that I only mask the five subunits (215 kD) in the center, which accounts for half of the filament in the box, to achieve a more accurate alignment. I’m not sure whether a mass of 215 kD would be enough to result in a good estimation of the per-particle scale.

Here is a screenshot of the cryosparc guide I mentioned above:

215kd should be fine for estimating per particle scale - that is a typical size for single particle analysis these days. You may well be right that using “none” is better in your case for other reasons, but I don’t think there will be any technical issue with refining scales in the preceding refinement.

1 Like