Contents
What can I do if autosol says "this version does not seem big enough"?
Autosol tries to automatically determine the size of solve or resolve, but if your data is very high resolution or a very large unit cell, you can get the message:
*************************************************** Sorry, this version does not seem big enough... (Current value of isizeit is 30) Unfortunately your computer will only accept a size of 30 with your current settings. You might try cutting back the resolution You might try "coarse_grid" to reduce memory You might try "unlimit" allow full use of memory ***************************************************
You cannot get rid of this problem by specifying the resolution with resolution=4.0 because autosol/autobuild/ligandfit uses the resolution cutoff you specify in all calculations, but the high-res data is still carried along.
The easiest solution to this problem is to edit your data file to have lower- resolution data. You can do it like this:
phenix.reflection_file_converter huge.sca --sca=big.sca --resolution=4.0
or in the GUI, use the reflection file editor.
A second solution is to tell autosol to ignore the high-res data explicitly with one of these commands (on the command line or in the GUI):
resolve_command="'resolution 200 4.0'" solve_command="'resolution 200 4.0'" resolve_pattern_command="'resolution 200 4.0'"
Note the two sets of quotes; both are required for this command-line input. Just one set of quotes is required in the GUI. These commands are applied after all other inputs in resolve/solve/resolve_pattern and therefore all data outside these limits will be ignored.
Why am I not allowed to use a file with FAVG SIGFAVG DANO SIGDANO in autosol or autobuild?
The group of MTZ columns FAVG SIGFAVG DANO SIGDANO is a special one that should normally not be used in Phenix. The reason is that Phenix stores this data as F+ SIGF+ F- SIGF-, but in the conversion process between F+/F- and FAVG/DANO, information is lost. Therefore you should normally supply data files with F+ SIGF+ F- SIGF- (or intensities), or fully merged data (F,SIG) to Phenix routines. As a special case, if you have anomalous data saved as FAVG SIGFAVG DANO SIGDANO you can supply this to autosol, however this requires either that (1) you supply a refinement file with F SIG, or that (2) your data file has a separate F SIG pair of columns (other than the FAVG SIGFAVG columns that are part of the FAVG/DANO group).
How can I specify a mask for density modification in autosol/autobuild/?
In autobuild you can simply use the command:
mask_from_pdb = my_mask_file.pdb rad_mask_from_pdb = 2
where my_mask_file.pdb has atoms in it marking the region to be masked. All points within rad_mask_from_pdb of an atom in my_mask_file.pdb will be considered inside the mask.
If you want to specify a mask in autosol, add this command:
resolve_command_list="'model ../../coords.pdb' 'use_model_mask'"
where there are " and ' quotes and coords.pdb is the model to use for a mask. Note the "../../" because coords.pdb is in your working directory but when resolve runs the run directory is 2 directories lower, so relative to that directory your coords.pdb is at "../../coords.pdb".
You will know it is working if your resolve_xx.log says:
Using model mask calculated from coordinates
Note: this command is most appropriate for use with the keyword maps_only=True because phenix.autobuild also uses model=... so that iterative model-building may not work entirely correctly in this case. Two parts that may not function correctly are "build_outside_model" (which will use your model as a mask and not the current one), and evaluate_model (which will evaluate your starting model, not the current model).
What do I do if autobuild says TRIED resolve_extra_huge ...but not OK?
In most cases when you get this error in phenix:
TRIED resolve_extra_huge ...but not OK
it actually means "your computer does not have enough memory to run resolve_extra_huge". If that is the case then you are kind of stuck unless you have another computer with even more memory+swap space, or you cut back on the resolution of the input data (Note that you have to actually lower the resolution in the input file, not just set "resolution=" because all the data is kept but not used if you just set the resolution).
You can also try the keyword:
resolve_command_list="'coarse_grid'"
(note 2 sets of quotes)
Sometimes the not OK message can happen if your system and PHENIX are not matching, so that resolve or solve cannot run at all. You can test for this by typing:
phenix.resolve
and if it loads up (just type QUIT or END or control-C to end it) then it runs, and if it doesn't, there is a system mismatch.
How can I tell autosol which columns to use from my mtz file?
In the GUI this is handled automatically, and available columns will be loaded into drop-down menus. On the command line, autosol will normally try to guess the appropriate columns of data from an input data file. If there are several choices, then you can tell autosol which one to use with the command_line keywords labels, peak.labels, infl.labels etc. For example if you have two input datafiles w1 and w2 for a 2-wavelength MAD dataset and you want to select the w1(+) and w1(-) data from the first file and w2(+) and w2(-1) from the second, you could use following keywords (see "How do I know what my choices of labels are for my data file" to know what to put in these lines):
input_file_list=" w1.mtz w2.mtz" group_labels_list=" 'w1(+) SIGw1(+) w1(-) SIGw1(-)' 'w2(+) SIGw2(+) w2(-) SIGw2(-)'"
Note that all the labels for one set of anomalous data from one file are grouped together in each set of quotes.
You could accomplish the same thing from a parameters file specifying something like:
wavelength{ wavelength_name = peak data = w1.mtz labels = w1(+) SIGw1(+) w1(-) SIGw1(-) } wavelength{ wavelength_name = infl data = w2.mtz labels = w2(+) SIGw2(+) w2(-) SIGw2(-) }
How do I know what my choices of labels are for my data file?
You can find out what your choices of labels are by running the command:
phenix.autosol show_labels=w1.mtz
This will provide a listing of the labels in w1.mtz and suggestions for their use in autosol/autobuild/ligandfit. For example the labels for w1.mtz yields:
List of all anomalous datasets in w1.mtz 'w1(+) SIGw1(+) w1(-) SIGw1(-)' List of all datasets in w1.mtz 'w1(+) SIGw1(+) w1(-) SIGw1(-)' List of all individual labels in w1.mtz 'w1(+)' 'SIGw1(+)' 'w1(-)' 'SIGw1(-)' Suggested uses: labels='w1(+) SIGw1(+) w1(-) SIGw1(-)' input_labels='w1(+) SIGw1(+) None None None None None None None' input_refinement_labels='w1(+) SIGw1(+) None' input_map_labels='w1(+) None None'
Why do I get "None of the solve versions worked" in autosol?
If you get this or a similar message for resolve, first have a look at LAST.LOG if it exists in your AutoSol_run_xx_ or AutoBuild_run_xx_ directory. The end of that file may give you a hint as to what was wrong.
The next thing to try is running one of these commands (just kill them with control-C if they do run):
phenix.solve
or:
phenix.resolve
If these load up solve or resolve, then they basically work and the problem is probably in the size of your dataset, some formatting issue, or the like.
If they do not run, then the problem is in your system setup. If you are using redhat linux, try changing the option of selinux to selinux=disabled in your /etc/sysconfig/selinux file.
It is also possible that you do not have the application csh installed on your system. If you have Ubuntu linux, csh and tcsh are not included in a normal installation. It is easy to install csh and tcsh under linux and it just takes a minute. On Ubuntu or Debian, you can say:
apt-get install tcsh
or on Fedora or CentOS and similar distributions:
yum install tcsh
and that should do it.
How can I do a quick check for iso and ano differences in an MIR dataset?
You can say:
phenix.autosol native.data=native.sca deriv.data=deriv.sca
and wait a couple minutes until it has scaled the data (once it says "RUNNING HYSS" you are far enough) and then have a look at:
AutoSol_run_1_/TEMP0/dataset_1_scale.log
which will say near the end:
isomorphous differences derivs 1 - native Differences by shell: shell dmin nobs Fbar R scale SIGNAL NOISE S/N 1 5.600 1018 285.012 0.287 0.998 105.05 26.73 3.93 2 4.200 1386 324.927 0.216 1.000 84.78 26.76 3.17 3 3.920 542 330.807 0.214 1.002 85.00 28.36 3.00 4 3.710 523 286.487 0.237 1.002 81.31 27.29 2.98 5 3.500 662 282.383 0.235 1.001 75.58 37.12 2.04 6 3.360 518 255.782 0.241 1.003 72.69 27.18 2.67 7 3.220 630 237.778 0.253 1.000 68.87 29.94 2.30 8 3.080 727 208.271 0.255 1.000 61.39 29.19 2.10 9 2.940 897 190.044 0.254 0.999 42.78 42.99 1.00 10 2.800 1067 169.022 0.280 0.999 50.54 33.24 1.52 Total: 7970 256.096 0.245 1.000 75.29 31.41 2.48
Here R is <Fderiv-Fnative>/(2 <Fderiv+Fnative>), noise is <sigma>, signal is sqrt(<(Fderiv-Fnative)**2>-<sigma**2>), and S/N is the ratio of signal to noise.
I ran autosol to get a partial model that I now want to refine. Which data file should I give as input: the original .sca file from HKL2000, or the file overall_best_refine_data.mtz from autosol?
Always use the MTZ file output by autosol. This contains a new set of R-free flags that have been used to refine the model; starting over with the .sca file will result in a new set of flags being generated, which biases R-free.
How can I calculate a log-likelihood gradient (LLG) map in Phaser?
See the maps section of the general FAQ list.