WAC

From Hall A Wiki
Revision as of 08:14, 30 November 2009 by Cjen (Talk | contribs) (WAC schedule)

(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to: navigation, search

The Weekly Analysis Coordinator (WAC) is a rotating position of 1 week each. The wac is required to perform their functions every morning. THIS INCLUDES WEEKENDS! The plots need to be up by NOON so that people can think about the results for the meeting.


WAC schedule

WAC schedule
Dates WAC name
2009/09/07 - 2009/09/13 ZAFAR AHMED
2009/09/14 - 2009/09/20 Luis Mercado
2009/09/21 - 2009/09/27 Xiaoyan Deng
2009/09/28 - 2009/10/04 Kai Pan
2009/10/05 - 2009/10/11 ZAFAR AHMED
2009/10/12 - 2009/10/18 Rupesh Silwal
2009/10/19 - 2009/10/26 ZAFAR AHMED
2009/11/04 - 2009/11/10 Xiaoyan Deng
2009/11/11 - 2009/11/17 Megan Friend
2009/11/18 - 2009/11/24 Luis Mercado
2009/11/25 - 2009/12/01 Kai Pan
2009/12/02 - 2009/12/08 Chun-Min (Mindy) Jen
2009/12/09 - 2009/12/15 Rupesh Silwal
2009/12/16 - 2009/12/23 Chun-Min (Mindy) Jen

Jobs of the WAC

Prompt the new runs

Run prompt on all the runs taken since the last wac summary. Do a reprompt if there is any doubt about the initial offline analysis. Look carefully through the output and try to spot problems. For example, Ahmed noticed large outliers in the bcm double differences and this helped us to learn that the dac glitch on the 18 bit ADCs was causing problems.

Update the run database

For every run the database of runs must be updated. The database structure is as follows.

+-------------------+------------+------+-----+-------------------+----------------+
| Field             | Type       | Null | Key | Default           | Extra          |
+-------------------+------------+------+-----+-------------------+----------------+
| WACentriesID      | int(11)    |      | PRI | NULL              | auto_increment | 
| RunInfo_RunNumber | int(11)    | YES  | MUL | NULL              |                | 
| WACname           | tinytext   | YES  |     | NULL              |                | 
| WACEntryTime      | timestamp  | YES  |     | CURRENT_TIMESTAMP |                | 
| WACcomment        | longtext   | YES  |     | NULL              |                | 
| LeftDETeventLo    | int(11)    | YES  |     | NULL              |                | 
| LeftDETeventHi    | int(11)    | YES  |     | NULL              |                | 
| RightDETeventLo   | int(11)    | YES  |     | NULL              |                | 
| RightDETeventHi   | int(11)    | YES  |     | NULL              |                | 
| Slug              | int(11)    | YES  |     | NULL              |                | 
| SlowSign          | tinyint(2) | YES  |     | NULL              |                | 
| Production        | tinyint(4) | YES  |     | NULL              |                | 
+-------------------+------------+------+-----+-------------------+----------------+

The database can be interacted with using the addWACentry.pl script in the wac directory:

addWACentry.pl -n [username] -r [runnumber]

Produce the slug summary

Once the database is up to date, the slug summary is produced using the summary.pl script in the wac directory. This will produce the summary root (~/pvdis/ROOTfiles/parity09_000X_slug.root) file and the summary text file (http://www.jlab.org/~adaq/pvdis/prompt_output/).

summary.pl -t slug -s [slugnumber]

Produce the slug plots

Once the slug summary has been produced, the slug plots are produced using the makePlots.pl script in the wac directory. This will produce plots at http://www.jlab.org/~adaq/pvdis/slug_output/

[apar@adaql4 ~]$ gowac

 ****** This directory for WAC analysis (WACs only, please) ******

[apar@adaql4 ~/pvdis/wac]$ 
[apar@adaql4 ~/pvdis/wac]$ makePlots.pl --type slug --slug 6 

Beware of clobbering, don't run two instances of makePlots.pl at the same time. They WILL clobber each other.

Special Cases

Dithering and regression slopes

If the runlist is up to date the slopes for the various detectors versus our readout bpm can be obtained as follows.

[apar@adaql6 ~/pvdis/wac]$ root
root [0] .x multirun/obtainBMslopes.C("dither")
root [1] .x multirun/obtainBMslopes.C("regress")

The may be complaints for runs if the dithering was off for that run or they were replayed without regression enabled. Once this is done, the plots are produced as follows.

[apar@adaql6 ~/pvdis/wac]$ ./utils/makeRegDitPlots.pl


Reprompting a lot of runs

On rare occasions it may be necessary to reprompt a whole bunch of runs. This can be accomplished in two steps.

1. First get a list of the runs to be reprompted. So, for example, all the runs in slug 5 can be obtained as follows.

[apar@adaql5 ~/pvdis/pan]$ gowac

 ****** This directory for WAC analysis (WACs only, please) ******

[apar@adaql5 ~/pvdis/wac]$ grep -P '\t5\t' runlist.txt | cut -f1 | xargs echo
12179 12180 12181 12182 12183 12184 12187 12189 12197 12198 12202 12203 12204 12207 12208 12209 12210 
12224 12225 12226 12227 12228 12234 12235 12236 12238 12240 12241 12242 12243 12244 12245 12246 12247 
12251 12253 12254 12255 12256 12257 12258 12259

2. Next use the following command.

[apar@adaql5 ~/pvdis/wac]$ gopan

 ****** This directory for ONLINE analysis (Shift-Workers only, please) ******

[apar@adaql5 ~/pvdis/pan]$ 
[apar@adaql5 ~/pvdis/pan]$ 
[apar@adaql5 ~/pvdis/pan]$ ./utils/prompt.pl  --forcepan --forcereg --forcedit --nomacro --conf '' --batch 
 12179 12180 12181 12182 12183 12184 12187 12189 12197 12198 12202 12203 12204 12207 12208 12209 12210 12224 
 12225 12226 12227 12228 12234 12235 12236 12238 12240 12241 12242 12243 12244 12245 12246 12247 12251 12253 
 12254 12255 12256 12257 12258 12259

Raw Data Files

from the adaq machines

The raw data files for runs were stored at
/adaql1/data2/parity/ 
but at 04:42:27 on September25,2009 Bob changed it so that new data goes to
/adaql1/data4/parity/ 
from run 12660 onwards.

from the silo

After some time these files are automatically removed. Once that happens, in order to view them you will have to retrieve them from the silo. This can be done as follows, from any ifarml machine.

ifarml5> jcache /mss/halla/parity/raw/parity09_12227.dat
[2009-10-06 10:29:36] Submitting request [11848406]
[2009-10-06 10:29:36] Request [11848406] is queued.
 

Some time after requesting the file, it will be available at

 /cache/mss/halla/parity/raw/parity09_12227.dat 

Backend Information

Software flowchart

Happex DAQ.gif

Choosing the data that goes into slug summaries

The data that goes into the root file output for the slug summary is controlled with the following file:

~/pvdis/wac/multirun/chooser.txt

The data that goes into the text file output is set with:

~/pvdis/wac/multirun/slug.C

Now, the text file output is generated from the root file, so, anything that appears in slug.C must also appear in chooser.txt for it to work.

Choosing what is plotted in the slug summary plots

Two files are used to control what is plot in the slug summaries. Pair tree variables are set up using the
~/pvdis/pan_cvs/pan/panguin/wac_slug.cfg
file. The epics variables are set up using the
~/pvdis/wac/epicsplot/doepicsplot.C
file. The latter is messy, but it should be reasonably clear how to change things. Make sure that any epics variable that are need appear in all of the following files.
/adaqfs/home/apar/scripts/fastEpicsLogger (Don't mess around with this file if you are not an expert. You can hang up the DAQ if you are not careful!!)
/adaqfs/home/apar/db/misc/Pvdis.misc
/adaqfs/home/apar/db/misc/PvdNoQweak.misc
After changes to either of these files, the
~/pvdis/wac/recent_slug_plots/index.html
file needs to be updated to reflect the correct number of pages that are produced as output. This is the file that eventually displays the plots as a web page.