/etc/pegasus/advanced.properties is in pegasus-wms 4.4.0+dfsg-7.
This file is owned by root:root, with mode 0o644.
The actual contents of the file can be viewed below.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 639 640 641 642 643 644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 776 777 778 779 780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 828 829 830 831 832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928 929 930 931 932 933 934 935 936 937 938 939 940 941 942 943 944 945 946 947 948 949 950 951 952 953 954 955 956 957 958 959 960 961 962 963 964 965 966 967 968 969 970 971 972 973 974 975 976 977 978 979 980 981 982 983 984 985 986 987 988 989 990 991 992 993 994 995 996 997 998 999 1000 1001 1002 1003 1004 1005 1006 1007 1008 1009 1010 1011 1012 1013 1014 1015 1016 1017 1018 1019 1020 1021 1022 1023 1024 1025 1026 1027 1028 1029 1030 1031 1032 1033 1034 1035 1036 1037 1038 1039 1040 1041 1042 1043 1044 1045 1046 1047 1048 1049 1050 1051 1052 1053 1054 1055 1056 1057 1058 1059 1060 1061 1062 1063 1064 1065 1066 1067 1068 1069 1070 1071 1072 1073 1074 1075 1076 1077 1078 1079 1080 1081 1082 1083 1084 1085 1086 1087 1088 1089 1090 1091 1092 1093 1094 1095 1096 1097 1098 1099 1100 1101 1102 1103 1104 1105 1106 1107 1108 1109 1110 1111 1112 1113 1114 1115 1116 1117 1118 1119 1120 1121 1122 1123 1124 1125 1126 1127 1128 1129 1130 1131 1132 1133 1134 1135 1136 1137 1138 1139 1140 1141 1142 1143 1144 1145 1146 1147 1148 1149 1150 1151 1152 1153 1154 1155 1156 1157 1158 1159 1160 1161 1162 1163 1164 1165 1166 1167 1168 1169 1170 1171 1172 1173 1174 1175 1176 1177 1178 1179 1180 1181 1182 1183 1184 1185 1186 1187 1188 1189 1190 1191 1192 1193 1194 1195 1196 1197 1198 1199 1200 1201 1202 1203 1204 1205 1206 1207 1208 1209 1210 1211 1212 1213 1214 1215 1216 1217 1218 1219 1220 1221 1222 1223 1224 1225 1226 1227 1228 1229 1230 1231 1232 1233 1234 1235 1236 1237 1238 1239 1240 1241 1242 1243 1244 1245 1246 1247 1248 1249 1250 1251 1252 1253 1254 1255 1256 1257 1258 1259 1260 1261 1262 1263 1264 1265 1266 1267 1268 1269 1270 1271 1272 1273 1274 1275 1276 1277 1278 1279 1280 1281 1282 1283 1284 1285 1286 1287 1288 1289 1290 1291 1292 1293 1294 1295 1296 1297 1298 1299 1300 1301 1302 1303 1304 1305 1306 1307 1308 1309 1310 1311 1312 1313 1314 1315 1316 1317 1318 1319 1320 1321 1322 1323 1324 1325 1326 1327 1328 1329 1330 1331 1332 1333 1334 1335 1336 1337 1338 1339 1340 1341 1342 1343 1344 1345 1346 1347 1348 1349 1350 1351 1352 1353 1354 1355 1356 1357 1358 1359 1360 1361 1362 1363 1364 1365 1366 1367 1368 1369 1370 1371 1372 1373 1374 1375 1376 1377 1378 1379 1380 1381 1382 1383 1384 1385 1386 1387 1388 1389 1390 1391 1392 1393 1394 1395 1396 1397 1398 1399 1400 1401 1402 1403 1404 1405 1406 1407 1408 1409 1410 1411 1412 1413 1414 1415 1416 1417 1418 1419 1420 1421 1422 1423 1424 1425 1426 1427 1428 1429 1430 1431 1432 1433 1434 1435 1436 1437 1438 1439 1440 1441 1442 1443 1444 1445 1446 1447 1448 1449 1450 1451 1452 1453 1454 1455 1456 1457 1458 1459 1460 1461 1462 1463 1464 1465 1466 1467 1468 1469 1470 1471 1472 1473 1474 1475 1476 1477 1478 1479 1480 1481 1482 1483 1484 1485 1486 1487 1488 1489 1490 1491 1492 1493 1494 1495 1496 1497 1498 1499 1500 1501 1502 1503 1504 1505 1506 1507 1508 1509 1510 1511 1512 1513 1514 1515 1516 1517 1518 1519 1520 1521 1522 1523 1524 1525 1526 1527 1528 1529 1530 1531 1532 1533 1534 1535 1536 1537 1538 1539 1540 1541 1542 1543 1544 1545 1546 1547 1548 1549 1550 1551 1552 1553 1554 1555 1556 1557 1558 1559 1560 1561 1562 1563 1564 1565 1566 1567 1568 1569 1570 1571 1572 1573 1574 1575 1576 1577 1578 1579 1580 1581 1582 1583 1584 1585 1586 1587 1588 1589 1590 1591 1592 1593 1594 1595 1596 1597 1598 1599 1600 1601 1602 1603 1604 1605 1606 1607 1608 1609 1610 1611 1612 1613 1614 1615 1616 1617 1618 1619 1620 1621 1622 1623 1624 1625 1626 1627 1628 1629 1630 1631 1632 1633 1634 1635 1636 1637 1638 1639 1640 1641 1642 1643 1644 1645 1646 1647 1648 1649 1650 1651 1652 1653 1654 1655 1656 1657 1658 1659 1660 1661 1662 1663 1664 1665 1666 1667 1668 1669 1670 1671 1672 1673 1674 1675 1676 1677 1678 1679 1680 1681 1682 1683 1684 1685 1686 1687 1688 1689 1690 1691 1692 1693 1694 1695 1696 1697 1698 1699 1700 1701 1702 1703 1704 1705 1706 1707 1708 1709 1710 1711 1712 1713 1714 1715 1716 1717 1718 1719 1720 1721 1722 1723 1724 1725 1726 1727 1728 1729 1730 1731 1732 1733 1734 1735 1736 1737 1738 1739 1740 1741 1742 1743 1744 1745 1746 1747 1748 1749 1750 1751 1752 1753 1754 1755 1756 1757 1758 1759 1760 1761 1762 1763 1764 1765 1766 1767 1768 1769 1770 1771 1772 1773 1774 1775 1776 1777 1778 1779 1780 1781 1782 1783 1784 1785 1786 1787 1788 1789 1790 1791 1792 1793 1794 1795 1796 1797 1798 1799 1800 1801 1802 1803 1804 1805 1806 1807 1808 1809 1810 1811 1812 1813 1814 1815 1816 1817 1818 1819 1820 1821 1822 1823 1824 1825 1826 1827 1828 1829 1830 1831 1832 1833 1834 1835 1836 1837 1838 1839 1840 1841 1842 1843 1844 1845 1846 1847 1848 1849 1850 1851 1852 1853 1854 1855 1856 1857 1858 1859 1860 1861 1862 1863 1864 1865 1866 1867 1868 1869 1870 1871 1872 1873 1874 1875 1876 1877 1878 1879 1880 1881 1882 1883 1884 1885 1886 1887 1888 1889 1890 1891 1892 1893 1894 1895 1896 1897 1898 1899 1900 1901 1902 1903 1904 1905 1906 1907 1908 1909 1910 1911 1912 1913 1914 1915 1916 1917 1918 1919 1920 1921 1922 1923 1924 1925 1926 1927 1928 1929 1930 1931 1932 1933 1934 1935 1936 1937 1938 1939 1940 1941 1942 1943 1944 1945 1946 1947 1948 1949 1950 1951 1952 1953 1954 1955 1956 1957 1958 1959 1960 1961 1962 1963 1964 1965 1966 1967 1968 1969 1970 1971 1972 1973 1974 1975 1976 1977 1978 1979 1980 1981 1982 1983 1984 1985 1986 1987 1988 1989 1990 1991 1992 1993 1994 1995 1996 1997 1998 1999 2000 2001 2002 2003 2004 2005 2006 2007 2008 2009 2010 2011 2012 2013 2014 2015 2016 2017 2018 2019 2020 2021 2022 2023 2024 2025 2026 2027 2028 2029 2030 2031 2032 2033 2034 2035 2036 2037 2038 2039 2040 2041 2042 2043 2044 2045 2046 2047 2048 2049 2050 2051 2052 2053 2054 2055 2056 2057 2058 2059 2060 2061 2062 2063 2064 2065 2066 2067 2068 2069 2070 2071 2072 2073 2074 2075 2076 2077 2078 2079 2080 2081 2082 2083 2084 2085 2086 2087 2088 2089 2090 2091 2092 2093 2094 2095 2096 2097 2098 2099 2100 2101 2102 2103 2104 2105 2106 2107 2108 2109 2110 2111 2112 2113 2114 2115 2116 2117 2118 2119 2120 2121 2122 2123 2124 2125 2126 2127 2128 2129 2130 2131 2132 2133 2134 2135 2136 2137 2138 2139 2140 2141 2142 2143 2144 2145 2146 2147 2148 2149 2150 2151 2152 2153 2154 2155 2156 2157 2158 2159 2160 2161 2162 2163 2164 2165 2166 2167 2168 2169 2170 2171 2172 2173 2174 2175 2176 2177 2178 2179 2180 2181 2182 2183 2184 2185 2186 2187 2188 2189 2190 2191 2192 2193 2194 2195 2196 2197 2198 2199 2200 2201 2202 2203 2204 2205 2206 2207 2208 2209 2210 2211 2212 2213 2214 2215 2216 2217 2218 2219 2220 2221 2222 2223 2224 2225 2226 2227 2228 2229 2230 2231 2232 2233 2234 2235 2236 2237 2238 2239 2240 2241 2242 2243 2244 2245 2246 2247 2248 2249 2250 2251 2252 2253 2254 2255 2256 2257 2258 2259 2260 2261 2262 2263 2264 2265 2266 2267 2268 2269 2270 2271 2272 2273 2274 2275 2276 2277 2278 2279 2280 2281 2282 2283 2284 2285 2286 2287 2288 2289 2290 2291 2292 2293 2294 2295 2296 2297 2298 2299 2300 2301 2302 2303 2304 2305 2306 2307 2308 2309 2310 2311 2312 2313 2314 2315 2316 2317 2318 2319 2320 2321 2322 2323 2324 2325 2326 2327 2328 2329 2330 2331 2332 2333 2334 2335 2336 2337 2338 2339 2340 2341 2342 2343 2344 2345 2346 2347 2348 2349 2350 2351 2352 2353 2354 2355 2356 2357 2358 2359 2360 2361 2362 2363 2364 2365 2366 2367 2368 2369 2370 2371 2372 2373 2374 2375 2376 2377 2378 2379 2380 2381 2382 2383 2384 2385 2386 2387 2388 2389 2390 2391 2392 2393 2394 2395 2396 2397 2398 2399 2400 2401 2402 2403 2404 2405 2406 2407 2408 2409 2410 2411 2412 2413 2414 2415 2416 2417 2418 2419 2420 2421 2422 2423 2424 2425 2426 2427 2428 2429 2430 2431 2432 2433 2434 2435 2436 2437 2438 2439 2440 2441 2442 2443 2444 2445 2446 2447 2448 2449 2450 2451 2452 2453 2454 2455 2456 2457 2458 2459 2460 2461 2462 2463 2464 2465 2466 2467 2468 2469 2470 2471 2472 2473 2474 2475 2476 2477 2478 2479 2480 2481 2482 2483 2484 2485 2486 2487 2488 2489 2490 2491 2492 2493 2494 2495 2496 2497 2498 2499 2500 2501 2502 2503 2504 2505 2506 2507 2508 2509 2510 2511 2512 2513 2514 2515 2516 2517 2518 2519 2520 2521 2522 2523 2524 2525 2526 2527 2528 2529 2530 2531 2532 2533 2534 2535 2536 2537 2538 2539 2540 2541 2542 2543 2544 2545 2546 2547 2548 2549 2550 2551 2552 2553 2554 2555 2556 2557 2558 2559 2560 2561 2562 2563 2564 2565 2566 2567 2568 2569 2570 2571 2572 2573 2574 2575 2576 2577 2578 2579 2580 2581 2582 2583 2584 2585 2586 2587 2588 2589 2590 2591 2592 2593 2594 2595 2596 2597 2598 2599 2600 2601 2602 2603 2604 2605 2606 2607 2608 2609 2610 2611 2612 2613 2614 2615 2616 2617 2618 2619 2620 2621 2622 2623 2624 2625 2626 2627 2628 2629 2630 2631 2632 2633 2634 2635 2636 2637 2638 2639 2640 2641 2642 2643 2644 2645 2646 2647 2648 2649 2650 2651 2652 2653 2654 2655 2656 2657 2658 2659 2660 2661 2662 2663 2664 2665 2666 2667 2668 2669 2670 2671 2672 2673 2674 2675 2676 2677 2678 2679 2680 2681 2682 2683 2684 2685 2686 2687 2688 2689 2690 2691 2692 2693 2694 2695 2696 2697 2698 2699 2700 2701 2702 2703 2704 2705 2706 2707 2708 2709 2710 2711 2712 2713 2714 2715 2716 2717 2718 2719 2720 2721 2722 2723 2724 2725 2726 2727 2728 2729 2730 2731 2732 2733 2734 2735 2736 2737 2738 2739 2740 2741 2742 2743 2744 2745 2746 2747 2748 2749 2750 2751 2752 2753 2754 2755 2756 2757 2758 2759 2760 2761 2762 2763 2764 2765 2766 2767 2768 2769 2770 2771 2772 2773 2774 2775 2776 2777 2778 2779 2780 2781 2782 2783 2784 2785 2786 2787 2788 2789 2790 2791 2792 2793 2794 2795 2796 2797 2798 2799 2800 2801 2802 2803 2804 2805 2806 2807 2808 2809 2810 2811 2812 2813 2814 2815 2816 2817 2818 2819 2820 2821 2822 2823 2824 2825 2826 2827 2828 2829 2830 2831 2832 2833 2834 2835 2836 2837 2838 2839 2840 2841 2842 2843 2844 2845 2846 2847 2848 2849 2850 2851 2852 2853 2854 2855 2856 2857 2858 2859 2860 2861 2862 2863 2864 2865 2866 2867 2868 2869 2870 2871 2872 2873 2874 2875 2876 2877 2878 2879 2880 2881 2882 2883 2884 2885 2886 2887 2888 2889 2890 2891 2892 2893 2894 2895 2896 2897 2898 2899 2900 2901 2902 2903 2904 2905 2906 2907 2908 2909 2910 2911 2912 2913 2914 2915 2916 2917 2918 2919 2920 2921 2922 2923 2924 2925 2926 2927 2928 2929 2930 2931 2932 2933 2934 2935 2936 2937 2938 2939 2940 2941 2942 2943 2944 2945 2946 2947 2948 2949 2950 2951 2952 2953 2954 2955 2956 2957 2958 2959 2960 2961 2962 2963 2964 2965 2966 2967 2968 2969 2970 2971 2972 2973 2974 2975 2976 2977 2978 2979 2980 2981 2982 2983 2984 2985 2986 2987 2988 2989 2990 2991 2992 2993 2994 2995 2996 2997 2998 2999 3000 3001 3002 3003 3004 3005 3006 3007 3008 3009 3010 3011 3012 3013 3014 3015 3016 3017 3018 3019 3020 3021 3022 3023 3024 3025 3026 3027 3028 3029 3030 3031 3032 3033 3034 3035 3036 3037 3038 3039 3040 3041 3042 3043 3044 3045 3046 3047 3048 3049 3050 3051 3052 3053 3054 3055 3056 3057 3058 3059 3060 3061 3062 3063 3064 3065 3066 3067 3068 3069 3070 3071 3072 3073 3074 3075 3076 3077 3078 3079 3080 3081 3082 3083 3084 3085 3086 3087 3088 3089 3090 3091 3092 3093 3094 3095 3096 3097 3098 3099 3100 3101 3102 3103 3104 3105 3106 3107 3108 3109 3110 3111 3112 3113 3114 3115 3116 3117 3118 3119 3120 3121 3122 3123 3124 3125 3126 3127 3128 3129 3130 3131 3132 3133 3134 3135 3136 3137 3138 3139 3140 3141 3142 3143 3144 3145 3146 3147 3148 3149 3150 3151 3152 3153 3154 3155 3156 3157 3158 3159 3160 3161 3162 3163 3164 3165 3166 3167 3168 3169 3170 3171 3172 3173 3174 3175 3176 3177 3178 3179 3180 3181 3182 3183 3184 3185 3186 3187 3188 3189 3190 3191 3192 3193 3194 3195 3196 3197 3198 3199 3200 3201 3202 3203 3204 3205 3206 3207 3208 3209 3210 3211 3212 3213 3214 3215 3216 3217 3218 3219 3220 3221 3222 3223 3224 3225 3226 3227 3228 3229 3230 3231 3232 3233 3234 3235 3236 3237 3238 3239 3240 3241 3242 | # TITLE "PROPERTIES"
#
# This is the reference guide to all properties regarding the
# Pegasus Workflow Planner, and their respective default values. Please refer
# to the user guide for a discussion when and which properties to use to
# configure various components. Please note that the values rely on
# proper capitalization, unless explicitly noted otherwise.
#
# Some properties rely with their default on the value of other
# properties. As a notation, the curly braces refer to the value of the
# named property. For instance, ${pegasus.home} means that the value depends
# on the value of the pegasus.home property plus any noted additions. You
# can use this notation to refer to other properties, though the extent
# of the subsitutions are limited. Usually, you want to refer to a set
# of the standard system properties. Nesting is not allowed.
# Substitutions will only be done once.
#
#
# There is a priority to the order of reading and evaluating properties.
# Usually one does not need to worry about the priorities. However, it
# is good to know the details of when which property applies, and how
# one property is able to overwrite another. The following is a mutually exclusive
# list ( highest priority first ) of property file locations.
#
# <orderedlist>
# <listitem>--conf option to the tools. Almost all of the clients that use properties
# have a --conf option to specify the property file to pick up.
# </listitem>
# <listitem> submit-dir/pegasus.xxxxxxx.properties file. All tools that work on the
# submit directory ( i.e after pegasus has planned a workflow) pick up the
# pegasus.xxxxx.properties file from the submit directory. The location for the
# pegasus.xxxxxxx.propertiesis picked up from the braindump file.
# </listitem>
# <listitem>The properties defined in the user property file
# <emphasis>${user.home}/.pegasusrc</emphasis> have lowest priority.
# </listitem>
# </orderedlist>
#
# Commandline properties have the highest priority. These override any property loaded
# from a property file. Each commandline property is introduced by a -D argument.
# Note that these arguments are parsed by the shell wrapper, and thus the -D arguments
# must be the first arguments to any command. Commandline properties are useful for debugging
# purposes.
#
# From Pegasus 3.1 release onwards, support has been dropped for the following
# properties that were used to signify the location of the properties file
#
# <itemizedlist>
# <listitem>pegasus.properties</listitem>
# <listitem>pegasus.user.properties</listitem>
# </itemizedlist>
#
# The following example provides a sensible set of properties to be set
# by the user property file. These properties use mostly non-default
# settings. It is an example only, and will not work for you:
#
# <screen>
# pegasus.catalog.replica File
# pegasus.catalog.replica.file ${pegasus.home}/etc/sample.rc.data
# pegasus.catalog.transformation Text
# pegasus.catalog.transformation.file ${pegasus.home}/etc/sample.tc.text
# pegasus.catalog.site.file ${pegasus.home}/etc/sample.sites.xml
# </screen>
#
# If you are in doubt which properties are actually visible, pegasus during the
# planning of the workflow dumps all properties after reading and prioritizing
# in the submit directory in a file with the suffix properties.
# Property : pegasus.home
# Systems : all
# Type : directory location string
# Default : "$PEGASUS_HOME"
#
# The property pegasus.home cannot be set in the property file. This property is
# automatically set up by the pegasus clients internally by determining the installation
# directory of pegasus. Knowledge about this property is important for developers who
# want to invoke PEGASUS JAVA classes without the shell wrappers.
#
# pegasus.home "$PEGASUS_HOME"
#
# SECTION "LOCAL DIRECTORIES"
#
# This section describes the GNU directory structure conventions. GNU
# distinguishes between architecture independent and thus sharable
# directories, and directories with data specific to a platform, and
# thus often local. It also distinguishes between frequently modified
# data and rarely changing data. These two axis form a space of four
# distinct directories.
# Property : pegasus.home.datadir
# Systems : all
# Type : directory location string
# Default : ${pegasus.home}/share
#
# The datadir directory contains broadly visiable and possilby exported
# configuration files that rarely change. This directory is currently
# unused.
#
# pegasus.home.datadir ${pegasus.home}/share
# Property : pegasus.home.sysconfdir
# Systems : all
# Type : directory location string
# Default : ${pegasus.home}/etc
#
# The system configuration directory contains configuration files that
# are specific to the machine or installation, and that rarely change.
# This is the directory where the XML schema definition copies are
# stored, and where the base pool configuration file is stored.
#
# pegasus.home.sysconfdir ${pegasus.home}/etc
# Property : pegasus.home.sharedstatedir
# Systems : all
# Type : directory location string
# Default : ${pegasus.home}/com
#
# Frequently changing files that are broadly visible are stored in the
# shared state directory. This is currently unused.
#
# pegasus.home.sharedstatedir ${pegasus.home}/com
# Property : pegasus.home.localstatedir
# Systems : all
# Type : directory location string
# Default : ${pegasus.home}/var
#
# Frequently changing files that are specific to a machine and/or
# installation are stored in the local state directory. This directory
# is being used for the textual transformation catalog,
# and the file-based replica catalog.
#
# pegasus.home.localstatedir ${pegasus.home}/var
# Property : pegasus.dir.submit.logs
# System : Pegasus
# Since : 2.4
# Type : directory location string
#
# This property can be used to specify the directory where the condor
# logs for the workflow should go to. By default, starting 4.2.1 release,
# Pegasus will setup the log to be in the workflow submit directory.
# This can create problems, in case users submit directories are on NSF.
#
# This is done to ensure that the logs are created in a local directory
# even though the submit directory maybe on NFS.
#
#
# pegasus.dir.submit.logs (no default)
#
# SECTION "SITE DIRECTORIES"
#
# The site directory properties modify the behavior of remotely run jobs.
# In rare occasions, it may also pertain to locally run compute jobs.
# Property : pegasus.dir.useTimestamp
# System : Pegasus
# Since : 2.1
# Type : Boolean
# Default : false
#
# While creating the submit directory, Pegasus employs a run numbering
# scheme. Users can use this property to use a timestamp based
# numbering scheme instead of the runxxxx scheme.
#
# pegasus.dir.useTimestamp false
# Property : pegasus.dir.exec
# System : Pegasus
# Since : 2.0
# Type : remote directory location string
# Default : (no default)
#
# This property modifies the remote location work directory in which all
# your jobs will run. If the path is relative then it is appended to the
# work directory (associated with the site), as specified in the site
# catalog. If the path is absolute then it overrides the work directory
# specified in the site catalog.
#
# pegasus.dir.exec (no default)
# Property : pegasus.dir.storage.mapper
# System : Pegasus
# Since : 4.3
# Type : enumeration
# Value[0] : Flat
# Value[1] : Fixed
# Value[2] : Hashed
# Value[3] : Replica
# Default : Flat
# See Also : pegasus.dir.storage.deep
#
# This property modifies determines how the output files are mapped on the
# output site storage location.
#
# In order to preserve backward compatibility, setting the boolean property
# pegasus.dir.storage.deep results in the Hashed output mapper to be loaded,
# if no output mapper property is specified.
#
# <variablelist>
# <varlistentry><term>Flat</term>
# <listitem>
# By default, Pegasus will place the output files in the storage directory
# specified in the site catalog for the output site.
# </listitem></varlistentry>
# <varlistentry><term>Fixed</term>
# <listitem>
# Using this mapper, users can specify an externally accesible url to
# the storage directory in their properties file. The following property
# needs to be set.
# <screen>
# pegasus.dir.storage.mapper.fixed.url an externally accessible URL to the
# storage directory on the output site
# e.g. gsiftp://outputs.isi.edu/shared/outputs
# </screen>
# Note: For hierarchal workflows, the above property needs to be set
# separately for each dax job, if you want the sub workflow outputs
# to goto a different directory.
# </listitem></varlistentry>
# <varlistentry><term>Hashed</term>
# <listitem>
# This mapper results in the creation of a deep directory structure
# on the output site, while populating the results. The base directory
# on the remote end is determined from the site catalog.
# Depending on the number of files being staged to the remote site a
# Hashed File Structure is created that ensures that only 256 files
# reside in one directory.
# To create this directory structure on the storage site, Pegasus
# relies on the directory creation feature of the Grid FTP server,
# which appeared in globus 4.0.x
# </listitem></varlistentry>
# <varlistentry><term>Replica</term>
# <listitem>
# This mapper determines the path for an output file on the output site by
# querying an output replica catalog. The output site is one that is
# passed on the command line. The output replica catalog can be configured
# by specifiing the properties with the prefix pegasus.dir.storage.replica.
# By default, a Regex File based backend is assumed unless overridden.
# For example
# <screen>
# pegasus.dir.storage.mapper.replica Regex|File
# pegasus.dir.storage.mapper.replica.file the RC file at the backend to use if using a file based RC
# </screen>
# </listitem></varlistentry>
# </variablelist>
#
# pegasus.dir.storage.mapper Flat
# Property : pegasus.dir.storage.deep
# System : Pegasus
# Since : 2.1
# Type : Boolean
# Default : false
# See Also : pegasus.dir.storage.mapper
#
#
# This property results in the creation of a deep directory structure
# on the output site, while populating the results. The base directory
# on the remote end is determined from the site catalog.
#
# To this base directory, the relative submit directory structure
# ( $user/$vogroup/$label/runxxxx ) is appended.
#
# $storage = $base + $relative_submit_directory
#
# This is the base directory that is passed to the storage mapper.
#
# Note: To preserve backward compatibilty, setting this
# property results in the Hashed mapper to be loaded unless
# pegasus.dir.storage.mapper is explicitly specified. Before 4.3,
# this property resulted in HashedDirectory structure.
#
# pegasus.dir.storage.deep false
# Property : pegasus.dir.create.strategy
# System : Pegasus
# Since : 2.2
# Type : enumeration
# Value[0] : HourGlass
# Value[1] : Tentacles
# Value[2] : Minimal
# Default : Minimal
#
# If the <screen>--randomdir</screen> option is given to the Planner at
# runtime, the Pegasus planner adds nodes that create the random
# directories at the remote pool sites, before any jobs are
# actually run. The two modes determine the placement of these
# nodes and their dependencies to the rest of the graph.
#
# <variablelist>
# <varlistentry><term>HourGlass</term>
# <listitem>
# It adds a make directory node at the top level of the graph, and all
# these concat to a single dummy job before branching out to the root
# nodes of the original/ concrete dag so far. So we introduce a
# classic X shape at the top of the graph. Hence the name HourGlass.
# </listitem></varlistentry>
# <varlistentry><term>Tentacles</term>
# <listitem>
# This option places the jobs creating directories at the top of the
# graph. However instead of constricting it to an hour glass shape,
# this mode links the top node to all the relevant nodes for which the
# create dir job is necessary. It looks as if the node spreads its
# tentacleas all around. This puts more load on the DAGMan because of
# the added dependencies but removes the restriction of the plan
# progressing only when all the create directory jobs have progressed
# on the remote pools, as is the case in the HourGlass model.
# </listitem></varlistentry>
# <varlistentry><term>Minimal</term>
# <listitem>
# The strategy involves in walking the graph in a BFS order, and
# updating a bit set associated with each job based on the BitSet
# of the parent jobs. The BitSet indicates whether an edge exists
# from the create dir job to an ancestor of the node.
# For a node, the bit set is the union of all the parents BitSets.
# The BFS traversal ensures that the bitsets are of a node are
# only updated once the parents have been processed.
# </listitem></varlistentry>
# </variablelist>
#
# pegasus.dir.create.strategy Tentacles
# Property : pegasus.dir.create.impl
# System : Pegasus
# Since : 2.2
# Type : enumeration
# Value[0] : DefaultImplementation
# Value[1] : S3
# Default : DefaultImpelmentation
#
# This property is used to select the executable that is used to
# create the working directory on the compute sites.
#
#
# <variablelist>
# <varlistentry><term>DefaultImplementation</term>
# <listitem>
# The default executable that is used to create a directory is the
# dirmanager executable shipped with Pegasus. It is found at
# $PEGASUS_HOME/bin/dirmanager in the pegasus distribution.
# An entry for transformation pegasus::dirmanager needs
# to exist in the Transformation Catalog or the PEGASUS_HOME
# environment variable should be specified in the site catalog for
# the sites for this mode to work.
# </listitem></varlistentry>
# <varlistentry><term>S3</term>
# <listitem>
# This option is used to create buckets in S3 instead of a
# directory. This should be set when running workflows on Amazon
# EC2. This implementation relies on s3cmd command line client to
# create the bucket. An entry for transformation amazon::s3cmd needs
# to exist in the Transformation Catalog for this to work.
# </listitem></varlistentry>
# </variablelist>
#
#
#
# pegasus.dir.create.impl DefaultImplementation
#
# SECTION "SCHEMA FILE LOCATION PROPERTIES"
#
# This section defines the location of XML schema files that are
# used to parse the various XML document instances in the PEGASUS. The
# schema backups in the installed file-system permit PEGASUS operations
# without being online.
# Property : pegasus.schema.dax
# Systems : Pegasus
# Since : 2.0
# Type : XML schema file location string
# Value[0] : ${pegasus.home.sysconfdir}/dax-3.2.xsd
# Default : ${pegasus.home.sysconfdir}/dax-3.2.xsd
#
# This file is a copy of the XML schema that describes abstract DAG
# files that are the result of the abstract planning process, and input
# into any concrete planning. Providing a copy of the schema enables the
# parser to use the local copy instead of reaching out to the internet,
# and obtaining the latest version from the GriPhyN website dynamically.
#
# pegasus.schema.dax ${pegasus.home.sysconfdir}/dax-3.2.xsd
# Property : pegasus.schema.sc
# Systems : Pegasus
# Since : 2.0
# Type : XML schema file location string
# Value[0] : ${pegasus.home.sysconfdir}/sc-3.0.xsd
# Default : ${pegasus.home.sysconfdir}/sc-3.0.xsd
#
# This file is a copy of the XML schema that describes the xml
# description of the site catalog, that is generated as a result of
# using genpoolconfig command.
# Providing a copy of the schema enables the parser to use the local
# copy instead of reaching out to the internet, and obtaining the
# latest version from the GriPhyN website dynamically.
#
# pegasus.schema.sc ${pegasus.home.sysconfdir}/sc-3.0.xsd
# Property : pegasus.schema.ivr
# Systems : all
# Type : XML schema file location string
# Value[0] : ${pegasus.home.sysconfdir}/iv-2.0.xsd
# Default : ${pegasus.home.sysconfdir}/iv-2.0.xsd
#
# This file is a copy of the XML schema that describes invocation record
# files that are the result of the a grid launch in a remote or local
# site. Providing a copy of the schema enables the parser to use the
# local copy instead of reaching out to the internet, and obtaining the
# latest version from the GriPhyN website dynamically.
#
# pegasus.schema.ivr ${pegasus.home.sysconfdir}/iv-2.1.xsd
#
# SECTION "DATABASE DRIVERS FOR ALL RELATIONAL CATALOGS"
#
# Property : pegasus.catalog.*.db.driver
# Property : pegasus.catalog.[catalog-name].db.driver
# System : Pegasus
# Type : Java class name
# Value[0] : Postgres
# Value[1] : MySQL
# Value[2] : SQLServer2000 (not yet implemented!)
# Value[3] : Oracle (not yet implemented!)
# Default : (no default)
# See also : pegasus.catalog.provenance
#
# The database driver class is dynamically loaded, as required by the
# schema. Currently, only PostGreSQL 7.3 and MySQL 4.0 are supported.
# Their respective JDBC3 driver is provided as part and parcel of the
# PEGASUS.
#
# A user may provide their own implementation, derived from
# org.griphyn.vdl.dbdriver.DatabaseDriver, to talk to a database of
# their choice.
#
# For each schema in PTC, a driver is instantiated
# separately, which has the same prefix as the schema. This may result
# in multiple connections to the database backend. As fallback, the
# schema "*" driver is attempted.
#
# The * in the property name can be replaced by a catalog name to
# apply the property only for that catalog.
# Valid catalog names are
#
# <screen>
# replica
# provenance
# </screen>
#
#
# pegasus.catalog.*.db.driver (no default)
# Property : pegasus.catalog.*.db.url
# Property : pegasus.catalog.[catalog-name].db.url
# System : PTC, ...
# Type : JDBC database URI string
# Default : (no default)
# Example : jdbc:postgresql:${user.name}
#
# Each database has its own string to contact the database on a given
# host, port, and database. Although most driver URLs allow to pass
# arbitrary arguments, please use the
# pegasus.catalog.[catalog-name].db.* keys or pegasus.catalog.*.db.*
# to preload these arguments.
# THE URL IS A MANDATORY PROPERTY FOR ANY DBMS BACKEND.
#
# <screen>
# Postgres : jdbc:postgresql:[//hostname[:port]/]database
# MySQL : jdbc:mysql://hostname[:port]]/database
# SQLServer: jdbc:microsoft:sqlserver://hostname:port
# Oracle : jdbc:oracle:thin:[user/password]@//host[:port]/service
# </screen>
#
# The * in the property name can be replaced by a catalog name to
# apply the property only for that catalog.
# Valid catalog names are
#
# <screen>
# replica
# provenance
# </screen>
#
#
#
# pegasus.catalog.*.db.url (no default)
# Property : pegasus.catalog.*.db.user
# Property : pegasus.catalog.[catalog-name].db.user
# System : PTC, ...
# Type : string
# Default : (no default)
# Example : ${user.name}
#
# In order to access a database, you must provide the name of your
# account on the DBMS. This property is database-independent. THIS IS A
# MANDATORY PROPERTY FOR MANY DBMS BACKENDS.
#
# The * in the property name can be replaced by a catalog name to
# apply the property only for that catalog.
# Valid catalog names are
#
# <screen>
# replica
# provenance
# </screen>
#
#
#
#
# pegasus.catalog.*.db.user (no default)
# Property : pegasus.catalog.*.db.password
# Property : pegasus.catalog.[catalog-name].db.password
# System : PTC, ...
# Type : string
# Default : (no default)
# Example : ${user.name}
#
# In order to access a database, you must provide an optional password
# of your account on the DBMS. This property is database-independent.
# THIS IS A MANDATORY PROPERTY, IF YOUR DBMS BACKEND ACCOUNT REQUIRES
# A PASSWORD.
#
# The * in the property name can be replaced by a catalog name to
# apply the property only for that catalog.
# Valid catalog names are
#
# <screen>
# replica
# provenance
# </screen>
#
#
#
# pegasus.catalog.*.db.password (no default)
# Property : pegasus.catalog.*.db.*
# Property : pegasus.catalog.[catalog-name].db.*
# System : PTC, RC
#
# Each database has a multitude of options to control in fine detail
# the further behaviour. You may want to check the JDBC3 documentation
# of the JDBC driver for your database for details. The keys will be
# passed as part of the connect properties by stripping the
# "pegasus.catalog.[catalog-name].db." prefix from them.
# The catalog-name can be replaced by the following values
# provenance for Provenance Catalog (PTC),
# replica for Replica Catalog (RC)
#
#
# Postgres 7.3 parses the following properties:
# <screen>
# pegasus.catalog.*.db.user
# pegasus.catalog.*.db.password
# pegasus.catalog.*.db.PGHOST
# pegasus.catalog.*.db.PGPORT
# pegasus.catalog.*.db.charSet
# pegasus.catalog.*.db.compatible
# </screen>
#
# MySQL 4.0 parses the following properties:
#
# <screen>
# pegasus.catalog.*.db.user
# pegasus.catalog.*.db.password
# pegasus.catalog.*.db.databaseName
# pegasus.catalog.*.db.serverName
# pegasus.catalog.*.db.portNumber
# pegasus.catalog.*.db.socketFactory
# pegasus.catalog.*.db.strictUpdates
# pegasus.catalog.*.db.ignoreNonTxTables
# pegasus.catalog.*.db.secondsBeforeRetryMaster
# pegasus.catalog.*.db.queriesBeforeRetryMaster
# pegasus.catalog.*.db.allowLoadLocalInfile
# pegasus.catalog.*.db.continueBatchOnError
# pegasus.catalog.*.db.pedantic
# pegasus.catalog.*.db.useStreamLengthsInPrepStmts
# pegasus.catalog.*.db.useTimezone
# pegasus.catalog.*.db.relaxAutoCommit
# pegasus.catalog.*.db.paranoid
# pegasus.catalog.*.db.autoReconnect
# pegasus.catalog.*.db.capitalizeTypeNames
# pegasus.catalog.*.db.ultraDevHack
# pegasus.catalog.*.db.strictFloatingPoint
# pegasus.catalog.*.db.useSSL
# pegasus.catalog.*.db.useCompression
# pegasus.catalog.*.db.socketTimeout
# pegasus.catalog.*.db.maxReconnects
# pegasus.catalog.*.db.initialTimeout
# pegasus.catalog.*.db.maxRows
# pegasus.catalog.*.db.useHostsInPrivileges
# pegasus.catalog.*.db.interactiveClient
# pegasus.catalog.*.db.useUnicode
# pegasus.catalog.*.db.characterEncoding
# </screen>
#
# MS SQL Server 2000 support the following properties (keys are
# case-insensitive, e.g. both "user" and "User" are valid):
#
# <screen>
# pegasus.catalog.*.db.User
# pegasus.catalog.*.db.Password
# pegasus.catalog.*.db.DatabaseName
# pegasus.catalog.*.db.ServerName
# pegasus.catalog.*.db.HostProcess
# pegasus.catalog.*.db.NetAddress
# pegasus.catalog.*.db.PortNumber
# pegasus.catalog.*.db.ProgramName
# pegasus.catalog.*.db.SendStringParametersAsUnicode
# pegasus.catalog.*.db.SelectMethod
# </screen>
#
# The * in the property name can be replaced by a catalog name to
# apply the property only for that catalog.
# Valid catalog names are
#
# <screen>
# replica
# provenance
# </screen>
#
#
#
# pegasus.catalog.*.db.* (no default)
#
# SECTION "CATALOG PROPERTIES"
#
#
# SUBSECTION "REPLICA CATALOG"
#
# Property : pegasus.catalog.replica
# System : Pegasus
# Since : 2.0
# Type : enumeration
# Value[0] : RLS
# Value[1] : LRC
# Value[2] : JDBCRC
# Value[3] : File
# Value[4] : Directory
# Value[5] : MRC
# Value[6] : Regex
# Default : RLS
#
# Pegasus queries a Replica Catalog to discover the physical filenames
# (PFN) for input files specified in the DAX. Pegasus can interface
# with various types of Replica Catalogs. This property specifies
# which type of Replica Catalog to use during the planning process.
#
# <variablelist>
# <varlistentry><term>RLS</term>
# <listitem> RLS (Replica Location Service) is a distributed replica
# catalog, which ships with GT4. There is an index service called
# Replica Location Index (RLI) to which 1 or more Local Replica
# Catalog (LRC) report. Each LRC can contain all or a subset of
# mappings. In this mode, Pegasus queries the central RLI to
# discover in which LRC's the mappings for a LFN reside. It then
# queries the individual LRC's for the PFN's.
# To use RLS, the user additionally needs to set the property
# pegasus.catalog.replica.url to specify the URL for the RLI to
# query.
# Details about RLS can be found at
# http://www.globus.org/toolkit/data/rls/
# </listitem></varlistentry>
# <varlistentry><term>LRC</term>
# <listitem> If the user does not want to query the RLI, but directly a
# single Local Replica Catalog.
# To use LRC, the user additionally needs to set the property
# pegasus.catalog.replica.url to specify the URL for the LRC to
# query.
# Details about RLS can be found at
# http://www.globus.org/toolkit/data/rls/
# </listitem></varlistentry>
# <varlistentry><term>JDBCRC</term>
# <listitem> In this mode, Pegasus queries a SQL based replica catalog that
# is accessed via JDBC. The sql schema's for this catalog can be
# found at $PEGASUS_HOME/sql directory.
# To use JDBCRC, the user additionally needs to set the following
# properties
# <orderedlist>
# <listitem>pegasus.catalog.replica.db.driver = mysql</listitem>
# <listitem>pegasus.catalog.replica.db.url = jdbc url to database e.g jdbc:mysql://database-host.isi.edu/database-name </listitem>
# <listitem>pegasus.catalog.replica.db.user = database-user</listitem>
# <listitem>pegasus.catalog.replica.db.password = database-password</listitem>
# </orderedlist>
# </listitem></varlistentry>
# <varlistentry><term>File</term>
# <listitem><para>In this mode, Pegasus queries a file based replica catalog.
# It is neither transactionally safe, nor advised to use for
# production purposes in any way. Multiple concurrent instances
# <emphasis>will clobber</emphasis> each other!. The site attribute should
# be specified whenever possible. The attribute key for the site
# attribute is "pool".
#
# The LFN may or may not be quoted. If it contains linear
# whitespace, quotes, backslash or an equality sign, it must be
# quoted and escaped. Ditto for the PFN. The attribute key-value
# pairs are separated by an equality sign without any
# whitespaces. The value may be in quoted. The LFN sentiments about quoting apply.
#
# <screen>
# LFN PFN
# LFN PFN a=b [..]
# LFN PFN a="b" [..]
# "LFN w/LWS" "PFN w/LWS" [..]
# </screen>
#
# To use File, the user additionally needs to specify
# pegasus.catalog.replica.file property to specify the path to the
# file based RC.
# </para>
# </listitem></varlistentry>
# <varlistentry>
# <term>Regex</term>
# <listitem><para>In this mode, Pegasus queries a file based replica catalog.
# It is neither transactionally safe, nor advised to use for
# production purposes in any way. Multiple concurrent access to
# the File will end up clobbering the contents of the file. The
# site attribute should be specified whenever possible. The attribute
# key for the site attribute is "pool".
#
# The LFN may or may not be quoted. If it contains linear
# whitespace, quotes, backslash or an equality sign, it must be
# quoted and escaped. Ditto for the PFN. The attribute key-value
# pairs are separated by an equality sign without any
# whitespaces. The value may be in quoted. The LFN sentiments about quoting apply.
#
# In addition users can specifiy regular expression based LFN's. A regular expression
# based entry should be qualified with an attribute named 'regex'. The attribute regex
# when set to true identifies the catalog entry as a regular expression based entry.
# Regular expressions should follow Java regular expression syntax.
#
# For example, consider a replica catalog as shown below.
#
# Entry 1 refers to an entry which does not use a resular expressions. This entry
# would only match a file named 'f.a', and nothing else.
# Entry 2 referes to an entry which uses a regular expression. In this entry f.a
# referes to files having name as f[any-character]a i.e. faa, f.a, f0a, etc.
#
# <screen>
# 1
# f.a file:///Volumes/data/input/f.a pool="local"
# 2
# f.a file:///Volumes/data/input/f.a pool="local" regex="true"
# </screen>
#
# Regular expression based entries also support substitutions. For example,
# consider the regular expression based entry shown below.
#
# Entry 3 will match files with name alpha.csv, alpha.txt, alpha.xml.
# In addition, values matched in the expression can be used to generate a PFN.
#
# For the entry below if the file being looked up is alpha.csv, the PFN for the file
# would be generated as file:///Volumes/data/input/csv/alpha.csv. Similary if the
# file being lookedup was alpha.csv, the PFN for the file would be generated as
# file:///Volumes/data/input/xml/alpha.xml i.e. The section [0], [1] will be replaced.
# Section [0] refers to the entire string i.e. alpha.csv. Section [1] refers to a partial
# match in the input i.e. csv, or txt, or xml. Users can utilize as many sections as they wish.
#
# 3
# <screen>
# alpha\.(csv|txt|xml) file:///Volumes/data/input/[1]/[0] pool="local" regex="true"
# </screen>
#
# To use File, the user additionally needs to specify
# pegasus.catalog.replica.file property to specify the path to the
# file based RC.
# </para></listitem>
# </varlistentry>
# <varlistentry><term>Directory</term>
# <listitem><para>In this mode, Pegasus does a directory listing on an input
# directory to create the LFN to PFN mappings. The directory listing is
# performed recursively, resulting in deep LFN mappings. For example, if an
# input directory $input is specified with the following structure
# <screen>
# $input
# $input/f.1
# $input/f.2
# $input/D1
# $input/D1/f.3
# </screen>
# Pegasus will create the mappings the following LFN PFN mappings internally
# <screen>
# f.1 file://$input/f.1 pool="local"
# f.2 file://$input/f.2 pool="local"
# D1/f.3 file://$input/D2/f.3 pool="local"
# </screen>
#
# If you don't want the deep lfn's to be created then, you can set
# pegasus.catalog.replica.directory.flat.lfn to true
# In that case, for the previous example, Pegasus will create the following
# LFN PFN mappings internally.
# <screen>
# f.1 file://$input/f.1 pool="local"
# f.2 file://$input/f.2 pool="local"
# f.3 file://$input/D2/f.3 pool="local"
# </screen>
#
# pegasus-plan has --input-dir option that can be used to specify an input
# directory.
#
# Users can optionally specify additional properties to configure the behvavior
# of this implementation.
#
# pegasus.catalog.replica.directory.site to specify a site attribute other than
# local to associate with the mappings.
#
# pegasus.catalog.replica.directory.url.prefix to associate a URL prefix for the PFN's
# constructed. If not specified, the URL defaults to file://
# </para>
# </listitem></varlistentry>
# <varlistentry><term>MRC</term>
# <listitem><para>In this mode, Pegasus queries multiple replica catalogs to
# discover the file locations on the grid. To use it set
#
# <screen>
# pegasus.catalog.replica MRC
# </screen>
#
# Each associated replica catalog can be configured via properties
# as follows.
#
# The user associates a variable name referred to as [value] for
# each of the catalogs, where [value] is any legal identifier
# (concretely [A-Za-z][_A-Za-z0-9]*) For each associated replica
# catalogs the user specifies the following properties.
#
# <screen>
# pegasus.catalog.replica.mrc.[value] specifies the type of replica catalog.
# pegasus.catalog.replica.mrc.[value].key specifies a property name key for a
# particular catalog
# </screen>
#
# For example, if a user wants to query two lrc's at the same time
# he/she can specify as follows
#
# <screen>
# pegasus.catalog.replica.mrc.lrc1 LRC
# pegasus.catalog.replica.mrc.lrc2.url rls://sukhna
#
# pegasus.catalog.replica.mrc.lrc2 LRC
# pegasus.catalog.replica.mrc.lrc2.url rls://smarty
# </screen>
#
#
# In the above example, lrc1, lrc2 are any valid identifier names
# and url is the property key that needed to be specified.
# </para>
# </listitem></varlistentry>
# </variablelist>
#
#
#
# pegasus.catalog.replica RLS
# Property : pegasus.catalog.replica.url
# System : Pegasus
# Since : 2.0
# Type : URI string
# Default : (no default)
#
# When using the modern RLS replica catalog, the URI to the Replica
# catalog must be provided to Pegasus to enable it to look up
# filenames. There is no default.
#
# pegasus.catalog.replica.url (no default)
# Property : pegasus.catalog.replica.chunk.size
# System : Pegasus, rc-client
# Since : 2.0
# Type : Integer
# Default : 1000
#
# The rc-client takes in an input file containing the mappings upon
# which to work. This property determines, the number of lines that
# are read in at a time, and worked upon at together. This allows the
# various operations like insert, delete happen in bulk if the
# underlying replica implementation supports it.
#
#
# pegasus.catalog.replica.chunk.size 1000
# Property : pegasus.catalog.replica.lrc.ignore
# System : Replica Catalog - RLS
# Since : 2.0
# Type : comma separated list of LRC urls
# Default : (no default)
# See also : pegasus.catalog.replica.lrc.restrict
#
# Certain users may like to skip some LRCs while querying for the physical
# locations of a file. If some LRCs need to be skipped from those found in the
# rli then use this property. You can define either the full URL or partial
# domain names that need to be skipped. E.g. If a user wants
# rls://smarty.isi.edu and all LRCs on usc.edu to be skipped then the
# property will be set as pegasus.rls.lrc.ignore=rls://smarty.isi.edu,usc.edu
#
# pegasus.catalog.replica.lrc.ignore (no default)
# Property : pegasus.catalog.replica.lrc.restrict
# System : Replica Catalog - RLS
# Since : 1.3.9
# Type : comma separated list of LRC urls
# Default : (no default)
# See also : pegasus.catalog.replica.lrc.ignore
#
# This property applies a tighter restriction on the results returned
# from the LRCs specified. Only those PFNs are returned that have a
# pool attribute associated with them. The property "pegasus.rc.lrc.ignore"
# has a higher priority than "pegasus.rc.lrc.restrict". For example, in case
# a LRC is specified in both properties, the LRC would be ignored (i.e.
# not queried at all instead of applying a tighter restriction on the
# results returned).
#
# pegasus.catalog.replica.lrc.restrict (no default)
# Property : pegasus.catalog.replica.lrc.site.[site-name]
# System : Replica Catalog - RLS
# Since : 2.3.0
# Type : LRC url
# Default : (no default)
#
#
# This property allows for the LRC url to be associated with site
# handles. Usually, a pool attribute is required to be associated with
# the PFN for Pegasus to figure out the site on which PFN resides.
# However, in the case where an LRC is responsible for only
# a single site's mappings, Pegasus can safely associate LRC url
# with the site. This association can be used to determine the pool
# attribute for all mappings returned from the LRC, if the mapping
# does not have a pool attribute associated with it.
#
# The site_name in the property should be replaced by the name of
# the site. For example
# <screen>
# pegasus.catalog.replica.lrc.site.isi rls://lrc.isi.edu
# </screen>
# tells Pegasus that all PFNs returned from LRC rls://lrc.isi.edu
# are associated with site isi.
#
# The [site_name] should be the same as the site handle specified in
# the site catalog.
#
# pegasus.catalog.replica.lrc.site.[site-name] (no default)
# Property : pegasus.catalog.replica.cache.asrc
# System : Pegasus
# Since : 2.0
# Type : Boolean
# Value[0] : false
# Value[1] : true
# Default : false
# See also : pegasus.catalog.replica
#
# This property determines whether to treat the cache file specified
# as a supplemental replica catalog or not. User can specify on the
# command line to pegasus-plan a comma separated list of cache files using
# the --cache option. By default, the LFN->PFN mappings contained in
# the cache file are treated as cache, i.e if an entry is found in a
# cache file the replica catalog is not queried. This results in only
# the entry specified in the cache file to be available for replica
# selection.
#
# Setting this property to true, results in the cache files to be
# treated as supplemental replica catalogs. This results in the
# mappings found in the replica catalog (as specified by
# pegasus.catalog.replica) to be merged with the ones found in the
# cache files. Thus, mappings for a particular LFN found in both the
# cache and the replica catalog are available for replica selection.
#
# pegasus.catalog.replica.cache.asrc false
#
# SUBSECTION "SITE CATALOG"
#
# Property : pegasus.catalog.site
# System : Site Catalog
# Since : 2.0
# Type : enumeration
# Value[0] : XML4
# Value[1] : XML3
# Default : XML4
#
# The site catalog file format is now automatically detected, so there
# should be no need to use the property anymore.
#
# pegasus.catalog.site XML4
# Property : pegasus.catalog.site.file
# System : Site Catalog
# Since : 2.0
# Type : file location string
# Default : ${pegasus.home.sysconfdir}/sites.xml
# See also : pegasus.catalog.site
#
# Running things on the grid requires an extensive description of the
# capabilities of each compute cluster, commonly termed "site". This
# property describes the location of the file that contains such a site
# description. As the format is currently in flow, please refer to the
# userguide and Pegasus for details which format is expected.
#
# pegasus.catalog.site.file ${pegasus.home.sysconfdir}/sites.xml
#
# SUBSECTION "TRANSFORMATION CATALOG"
#
# Property : pegasus.catalog.transformation
# System : Transformation Catalog
# Since : 2.0
# Type : enumeration
# Value[0] : Text
# Value[1] : File
# Default : Text
# See also : pegasus.catalog.transformation.file
#
#
# <variablelist>
# <varlistentry><term>Text</term>
# <listitem><para>In this mode, a multiline file based format is understood. The file
# is read and cached in memory. Any modifications, as adding or
# deleting, causes an update of the memory and hence to the file
# underneath. All queries are done against the memory
# representation.
#
# The file sample.tc.text in the etc directory contains an example
#
# Here is a sample textual format for transfomation catalog containing
# one transformation on two sites
#
# <screen>
# tr example::keg:1.0 {
#
# #specify profiles that apply for all the sites for the transformation
# #in each site entry the profile can be overriden
# profile env "APP_HOME" "/tmp/karan"
# profile env "JAVA_HOME" "/bin/app"
#
# site isi {
# profile env "me" "with"
# profile condor "more" "test"
# profile env "JAVA_HOME" "/bin/java.1.6"
# pfn "/path/to/keg"
# arch "x86"
# os "linux"
# osrelease "fc"
# osversion "4"
# type "INSTALLED"
# }
#
# site wind {
# profile env "me" "with"
# profile condor "more" "test"
# pfn "/path/to/keg"
# arch "x86"
# os "linux"
# osrelease "fc"
# osversion "4"
# type "STAGEABLE"
# }
# }
# </screen>
# </para>
# </listitem></varlistentry>
# <varlistentry><term>File</term>
# <listitem>THIS FORMAT IS DEPRECATED. WILL BE REMOVED IN COMING VERSIONS.
# USE pegasus-tc-converter to convert File format to Text Format.
# In this mode, a file format is understood. The file is
# read and cached in memory. Any modifications, as adding or
# deleting, causes an update of the memory and hence to the file
# underneath. All queries are done against the memory
# representation. The new TC file format uses 6 columns:
# <orderedlist>
# <listitem>The resource ID is represented in the first column.</listitem>
# <listitem>The logical transformation uses the colonized format
# ns::name:vs.</listitem>
# <listitem>The path to the application on the system</listitem>
# <listitem>The installation type is identified by one of the following
# keywords - all upper case: INSTALLED, STAGEABLE.
# If not specified, or <command>NULL</command> is used, the type
# defaults to INSTALLED.</listitem>
# <listitem>The system is of the format ARCH::OS[:VER:GLIBC]. The
# following arch types are understood: "INTEL32", "INTEL64",
# "SPARCV7", "SPARCV9".
# The following os types are understood: "LINUX", "SUNOS",
# "AIX". If unset or <command>NULL</command>, defaults to
# INTEL32::LINUX.</listitem>
# <listitem>Profiles are written in the format
# NS::KEY=VALUE,KEY2=VALUE;NS2::KEY3=VALUE3
# Multiple key-values for same namespace are seperated by a
# comma "," and multiple namespaces are seperated by a
# semicolon ";". If any of your profile values contains a
# comma you must not use the namespace abbreviator.</listitem>
# </orderedlist>
# </listitem></varlistentry>
# </variablelist>
#
#
# pegasus.catalog.transformation Text
# Property : pegasus.catalog.transformation.file
# Systems : Transformation Catalog
# Type : file location string
# Default : ${pegasus.home.sysconfdir}/tc.text | ${pegasus.home.sysconfdir}/tc.data
# See also : pegasus.catalog.transformation
#
# This property is used to set the path to the textual transformation
# catalogs of type File or Text. If the transformation catalog is of type Text
# then tc.text file is picked up from sysconfdir, else tc.data
#
#
# pegasus.catalog.transformation.file ${pegasus.home.sysconfdir}/tc.text | ${pegasus.home.sysconfdir}/tc.data
#
# SUBSECTION "PROVENANCE CATALOG"
#
# Property : pegasus.catalog.provenance
# System : Provenance Tracking Catalog (PTC)
# Since : 2.0
# Type : Java class name
# Value[0] : InvocationSchema
# Value[1] : NXDInvSchema
# Default : (no default)
# See also : pegasus.catalog.*.db.driver
#
# This property denotes the schema that is being used to access a PTC.
# The PTC is usually not a standard installation. If you use a database
# backend, you most likely have a schema that supports PTCs. By default,
# no PTC will be used.
#
# Currently only the InvocationSchema is available for storing the
# provenance tracking records. Beware, this can become a lot of data.
# The values are names of Java classes. If no absolute Java classname
# is given, "org.griphyn.vdl.dbschema." is prepended. Thus, by deriving
# from the DatabaseSchema API, and implementing the PTC interface,
# users can provide their own classes here.
#
# Alternatively, if you use a native XML database like eXist, you can
# store data using the NXDInvSchema. This will avoid using any of the
# other database driver properties.
#
# pegasus.catalog.provenance (no default)
# Property : pegasus.catalog.provenance.refinement
# System : PASOA Provenance Store
# Since : 2.0.1
# Type : Java class name
# Value[0] : Pasoa
# Value[1] : InMemory
# Default : InMemory
# See also : pegasus.catalog.*.db.driver
#
# This property turns on the logging of the refinement process that
# happens inside Pegasus to the PASOA store. Not all actions are
# currently captured. It is still an experimental feature.
#
# The PASOA store needs to run on localhost on port 8080
# https://localhost:8080/prserv-1.0
#
#
# pegasus.catalog.provenance.refinement InMemory
#
# SECTION "REPLICA SELECTION PROPERTIES"
#
# Property : pegasus.selector.replica
# System : Replica Selection
# Since : 2.0
# Type : URI string
# Default : default
# See also : pegasus.replica.*.ignore.stagein.sites
# See also : pegasus.replica.*.prefer.stagein.sites
#
# Each job in the DAX maybe associated with input LFN's denoting the
# files that are required for the job to run. To determine the
# physical replica (PFN) for a LFN, Pegasus queries the replica
# catalog to get all the PFN's (replicas) associated with a LFN.
# Pegasus then calls out to a replica selector to select a replica
# amongst the various replicas returned. This property determines the
# replica selector to use for selecting the replicas.
#
# <variablelist>
# <varlistentry><term>Default</term>
# <listitem>
# If a PFN that is a file URL (starting with file:///) and has a
# pool attribute matching to the site handle of the site where the
# compute is to be run is found, then that is returned.
# Else,a random PFN is selected amongst all the PFN's that
# have a pool attribute matching to the site handle of the site
# where a compute job is to be run.
# Else, a random pfn is selected amongst all the PFN's.
# </listitem></varlistentry>
# <varlistentry><term>Restricted</term>
# <listitem><para>
# This replica selector, allows the user to specify good sites and
# bad sites for staging in data to a particular compute site. A good
# site for a compute site X, is a preferred site from which
# replicas should be staged to site X. If there are more than one
# good sites having a particular replica, then a random site is
# selected amongst these preferred sites.
#
# A bad site for a compute site X, is a site from which replica's
# should not be staged. The reason of not accessing replica from a
# bad site can vary from the link being down, to the user not having
# permissions on that site's data.
#
# The good | bad sites are specified by the properties
#
# <screen>
# pegasus.replica.*.prefer.stagein.sites
# pegasus.replica.*.ignore.stagein.sites
# </screen>
#
# where the * in the property name denotes the name of the compute
# site. A * in the property key is taken to mean all sites.
#
# The pegasus.replica.*.prefer.stagein.sites property takes precedence
# over pegasus.replica.*.ignore.stagein.sites property i.e. if for a
# site X, a site Y is specified both in the ignored and the
# preferred set, then site Y is taken to mean as only a preferred
# site for a site X.
# </para>
# </listitem></varlistentry>
# <varlistentry><term>Regex</term>
# <listitem><para>
# This replica selector allows the user allows the user to specific regex
# expressions that can be used to rank various PFN's returned from the
# Replica Catalog for a particular LFN. This replica selector selects the
# highest ranked PFN i.e the replica with the lowest rank value.
#
# The regular expressions are assigned different rank, that determine
# the order in which the expressions are employed. The rank values for
# the regex can expressed in user properties using the property.
#
# <screen>
# pegasus.selector.replica.regex.rank.[value] regex-expression
# </screen>
#
# The value is an integer value that denotes the rank of an expression
# with a rank value of 1 being the highest rank.
#
# Please note that before applying any regular expressions on
# the PFN's, the file URL's that dont match the preferred site are
# explicitly filtered out.
# </para>
# </listitem></varlistentry>
# <varlistentry><term>Local</term>
# <listitem>
# This replica selector prefers replicas from the local host and that
# start with a file: URL scheme. It is useful, when users want to
# stagin files to a remote site from your submit host using the
# Condor file transfer mechanism.
# </listitem></varlistentry>
# </variablelist>
#
#
# pegasus.selector.replica Default
# Property : pegasus.selector.replica.*.ignore.stagein.sites
# System : Replica Selection
# Type : comma separated list of sites
# Since : 2.0
# Default : no default
# See also : pegasus.selector.replica
# See also : pegasus.selector.replica.*.prefer.stagein.sites
#
# A comma separated list of storage sites from which to never stage in
# data to a compute site. The property can apply to all or a single
# compute site, depending on how the * in the property name is expanded.
#
# The * in the property name means all compute sites unless replaced
# by a site name.
#
# For e.g setting pegasus.selector.replica.*.ignore.stagein.sites to usc means that
# ignore all replicas from site usc for staging in to any compute site.
# Setting pegasus.replica.isi.ignore.stagein.sites to usc means that
# ignore all replicas from site usc for staging in data to site isi.
#
#
#
# pegasus.selector.replica.*.ignore.stagein.sites (no default)
# Property : pegasus.selector.replica.*.prefer.stagein.sites
# System : Replica Selection
# Type : comma separated list of sites
# Since : 2.0
# Default : no default
# See also : pegasus.selector.replica
# See also : pegasus.selector.replica.*.ignore.stagein.sites
#
# A comma separated list of preferred storage sites from which to stage in
# data to a compute site. The property can apply to all or a single
# compute site, depending on how the * in the property name is expanded.
#
# The * in the property name means all compute sites unless replaced
# by a site name.
#
# For e.g setting pegasus.selector.replica.*.prefer.stagein.sites to usc means that
# prefer all replicas from site usc for staging in to any compute site.
# Setting pegasus.replica.isi.prefer.stagein.sites to usc means that
# prefer all replicas from site usc for staging in data to site isi.
#
#
#
# pegasus.selector.replica.*.ignore.stagein.sites (no default)
# Property : pegasus.selector.replica.regex.rank.[value]
# System : Replica Selection
# Type : Regex Expression
# Since : 2.3.0
# Default : no default
# See also : pegasus.selector.replica
#
#
# Specifies the regex expressions to be applied on the PFNs returned
# for a particular LFN. Refer to
# <screen>
# http://java.sun.com/javase/6/docs/api/java/util/regex/Pattern.html
# </screen>
# on information of how to construct a regex expression.
#
# The [value] in the property key is to be replaced by an int value
# that designates the rank value for the regex expression to be
# applied in the Regex replica selector.
#
# The example below indicates preference for file URL's over
# URL's referring to gridftp server at example.isi.edu
#
# <screen>
# pegasus.selector.replica.regex.rank.1 file://.*
# pegasus.selector.replica.regex.rank.2 gsiftp://example\.isi\.edu.*
# </screen>
#
#
# pegasus.selector.replica.regex.rank.[value] regex-expression
#
# SECTION "SITE SELECTION PROPERTIES"
#
# Property : pegasus.selector.site
# System : Pegasus
# Since : 2.0
# Type : enumeration
# Value[0] : Random
# Value[1] : RoundRobin
# Value[2] : NonJavaCallout
# Value[3] : Group
# Value[4] : Heft
# Default : Random
# See also : pegasus.selector.site.path
# See also : pegasus.selector.site.timeout
# See also : pegasus.selector.site.keep.tmp
# See also : pegasus.selector.site.env.*
#
# The site selection in Pegasus can be on basis of any of the
# following strategies.
#
# <variablelist>
# <varlistentry><term>Random</term>
# <listitem>In this mode, the jobs will be randomly distributed among the
# sites that can execute them.
# </listitem></varlistentry>
# <varlistentry><term>RoundRobin</term>
# <listitem>In this mode. the jobs will be assigned in a round
# robin manner amongst the sites that can execute them. Since
# each site cannot execute everytype of job, the round robin
# scheduling is done per level on a sorted list. The sorting is
# on the basis of the number of jobs a particular site has been
# assigned in that level so far. If a job cannot be run on the
# first site in the queue (due to no matching entry in the
# transformation catalog for the transformation referred to by
# the job), it goes to the next one and so on. This implementation
# defaults to classic round robin in the case where all the jobs
# in the workflow can run on all the sites.
# </listitem></varlistentry>
# <varlistentry><term>NonJavaCallout</term>
# <listitem><para>In this mode, Pegasus will callout to an external site
# selector.In this mode a temporary file is prepared containing
# the job information that is passed to the site selector as an
# argument while invoking it. The path to the site selector is
# specified by setting the property pegasus.site.selector.path. The
# environment variables that need to be set to run the site
# selector can be specified using the properties with a
# pegasus.site.selector.env. prefix. The temporary file contains
# information about the job that needs to be scheduled. It
# contains key value pairs with each key value pair being on a
# new line and separated by a =.
#
# The following pairs are currently generated for the site
# selector temporary file that is generated in the NonJavaCallout.
#
# <tb2>
# version & is the version of the site selector
# api,currently 2.0.</te>
# transformation & is the fully-qualified definition
# identifier for the transformation (TR)
# namespace::name:version. </te>
# derivation & is teh fully qualified definition
# identifier for the derivation (DV),
# namespace::name:version. </te>
# job.level & is the job's depth in the tree of the
# workflow DAG. </te>
# job.id & is the job's ID, as used in the DAX
# file. </te>
# resource.id & is a pool handle, followed by whitespace,
# followed by a gridftp server. Typically,
# each gridftp server is enumerated once,
# so you may have multiple occurances of
# the same site. There can be multiple
# occurances of this key. </te>
# input.lfn & is an input LFN, optionally followed by a
# whitespace and file size. There can be
# multiple occurances of this key,one for
# each input LFN required by the job.</te>
# wf.name & label of the dax, as found in the DAX's
# root element.
# wf.index is the DAX index, that is incremented for
# each partition in case of deferred
# planning.</te>
# wf.time & is the mtime of the workflow. </te>
# wf.manager & is the name of the workflow manager being
# used .e.g condor </te>
# vo.name & is the name of the virtual organization
# that is running this workflow. It is
# currently set to NONE </te>
# vo.group & unused at present and is set to NONE. </te>
# </tb2>
# </para>
# </listitem></varlistentry>
# <varlistentry><term>Group</term>
# <listitem>In this mode, a group of jobs will be assigned to the same
# site that can execute them. The use of the PEGASUS profile key
# group in the dax, associates a job with a particular group. The
# jobs that do not have the profile key associated with them,
# will be put in the default group. The jobs in the
# default group are handed over to the "Random" Site Selector for
# scheduling.
# </listitem></varlistentry>
# <varlistentry><term>Heft</term>
# <listitem><para>In this mode, a version of the HEFT processor scheduling
# algorithm is used to schedule jobs in the workflow to multiple
# grid sites. The implementation assumes default data
# communication costs when jobs are not scheduled on to the same
# site. Later on this may be made more configurable.
#
# The runtime for the jobs is specified in the transformation
# catalog by associating the pegasus profile key runtime with the
# entries.
#
# The number of processors in a site is picked up from the
# attribute idle-nodes associated with the vanilla jobmanager of
# the site in the site catalog.
# </para>
# </listitem></varlistentry>
# </variablelist>
#
# pegasus.selector.site Random
# Property : pegasus.selector.site.path
# System : Site Selector
# Since : 2.0
# Type : String
#
# If one calls out to an external site selector using the
# NonJavaCallout mode, this refers to the path where the site selector
# is installed. In case other strategies are used it does not need to
# be set.
#
# pegasus.site.selector.path (no default)
# Property : pegasus.site.selector.env.*
# System : Pegasus
# Since : 1.2.3
# Type : String
#
# The environment variables that need to be set while callout to the
# site selector. These are the variables that the user would set if
# running the site selector on the command line. The name of the
# environment variable is got by stripping the keys of the prefix
# "pegasus.site.selector.env." prefix from them. The value of the
# environment variable is the value of the property.
#
# e.g pegasus.site.selector.path.LD_LIBRARY_PATH /globus/lib would lead to
# the site selector being called with the LD_LIBRARY_PATH set to
# /globus/lib.
#
# pegasus.site.selector.env.* (no default)
# Property : pegasus.selector.site.timeout
# System : Site Selector
# Since : 2.0
# Type : non negative integer
# Default : 60
#
# It sets the number of seconds Pegasus waits to hear back from an
# external site selector using the NonJavaCallout interface before
# timing out.
#
# pegasus.selector.site.timeout 60
# Property : pegasus.selector.site.keep.tmp
# System : Pegasus
# Since : 2.0
# Type : enumeration
# Value[0] : onerror
# Value[1] : always
# Value[2] : never
# Default : onerror
#
# It determines whether Pegasus deletes the temporary input files that
# are generated in the temp directory or not. These temporary input
# files are passed as input to the external site selectors.
#
# A temporary input file is created for each that needs to be scheduled.
#
# pegasus.selector.site.keep.tmp onerror
#
# SECTION "DATA STAGING CONFIGURATION"
#
# Property : pegasus.data.configuration
# System : Pegasus
# Since : 4.0
# Type : enumeration
# Value[0] : sharedfs
# Value[1] : nonsharedfs
# Value[2] : condorio
# Default : sharedfs
#
# This property sets up Pegasus to run in different environments.
#
#
# <variablelist>
# <varlistentry><term>sharedfs</term>
# <listitem>If this is set, Pegasus will be setup to execute jobs on the shared
# filesystem on the execution site. This assumes, that the head node of a cluster
# and the worker nodes share a filesystem. The staging site in this case is
# the same as the execution site. Pegasus adds a create dir job to the executable
# workflow that creates a workflow specific directory on the shared filesystem .
# The data transfer jobs in the executable workflow ( stage_in_ , stage_inter_ ,
# stage_out_ ) transfer the data to this directory.The compute jobs in the
# executable workflow are launched in the directory on the shared filesystem.
# Internally, if this is set the following properties are set.
# <screen>
# pegasus.execute.*.filesystem.local false
# </screen>
# </listitem>
# </varlistentry>
# <varlistentry><term>condorio</term>
# <listitem>If this is set, Pegasus will be setup to run jobs in a pure condor pool,
# with the nodes not sharing a filesystem. Data is staged to the compute nodes from
# the submit host using Condor File IO.
# The planner is automatically setup to use the submit host ( site local ) as the
# staging site. All the auxillary jobs added by the planner to the executable
# workflow ( create dir, data stagein and stage-out, cleanup ) jobs refer to
# the workflow specific directory on the local site. The data transfer jobs in
# the executable workflow ( stage_in_ , stage_inter_ , stage_out_ ) transfer the
# data to this directory. When the compute jobs start, the input data for each
# job is shipped from the workflow specific directory on the submit host to
# compute/worker node using Condor file IO. The output data for each job is
# similarly shipped back to the submit host from the compute/worker node.
# This setup is particularly helpful when running workflows in the cloud
# environment where setting up a shared filesystem across the VM's may be
# tricky.
# On loading this property, internally the following properies are set
# <screen>
# pegasus.transfer.lite.*.impl Condor
# pegasus.execute.*.filesystem.local true
# pegasus.gridstart PegasusLite
# pegasus.transfer.worker.package true
# </screen>
# </listitem>
# </varlistentry>
# <varlistentry><term>nonsharedfs</term>
# <listitem>If this is set, Pegasus will be setup to execute jobs on an execution site
# without relying on a shared filesystem between the head node and the worker nodes.
# You can specify staging site ( using --staging-site option to pegasus-plan) to
# indicate the site to use as a central storage location for a workflow. The
# staging site is independant of the execution sites on which a workflow executes.
# All the auxillary jobs added by the planner to the executable
# workflow ( create dir, data stagein and stage-out, cleanup ) jobs refer to
# the workflow specific directory on the staging site. The data transfer jobs in
# the executable workflow ( stage_in_ , stage_inter_ , stage_out_ ) transfer the
# data to this directory. When the compute jobs start, the input data for each
# job is shipped from the workflow specific directory on the submit host to
# compute/worker node using pegasus-transfer. The output data for each job is
# similarly shipped back to the submit host from the compute/worker node.
# The protocols supported are at this time SRM, GridFTP, iRods, S3.
# This setup is particularly helpful when running workflows on OSG where
# most of the execution sites don't have enough data storage. Only a few
# sites have large amounts of data storage exposed that can be used to place
# data during a workflow run. This setup is also helpful when running workflows
# in the cloud environment where setting up a shared filesystem across the VM's may be
# tricky.
# On loading this property, internally the following properies are set
# <screen>
# pegasus.execute.*.filesystem.local true
# pegasus.gridstart PegasusLite
# pegasus.transfer.worker.package true
# </screen>
# </listitem></varlistentry>
# </variablelist>
#
#
#
# pegasus.data.configuration sharedfs
# Property : pegasus.transfer.bypass.input.staging
# System : Pegasus
# Since : 4.3
# Type : Boolean
# Default : (no default)
# See also : pegasus.data.configuration
#
# When executiing in a non shared filesystem setup i.e data configuration set to nonsharedfs
# or condorio, Pegasus always stages the input files through the staging site i.e the stage-in
# job stages in data from the input site to the staging site. The PegasusLite jobs that start
# up on the worker nodes, then pull the input data from the staging site for each job.
#
# This property can be used to setup the PegasusLite jobs to pull input data directly
# from the input site without going through the staging server. This is based on the
# assumption that the worker nodes can access the input site. If users set this to true,
# they should be aware that the access to the input site is no longer throttled ( as in case
# of stage in jobs). If large number of compute jobs start at the same time in a workflow,
# the input server will see a connection from each job.
#
#
#
# pegasus.transfer.bypass.input.staging false
#
# SECTION "TRANSFER CONFIGURATION PROPERTIES"
#
# Property : pegasus.transfer.*.impl
# System : Pegasus
# Type : enumeration
# Value[0] : Transfer
# Value[1] : GUC
# Default : Transfer
# See also : pegasus.transfer.refiner
# Since : 2.0
#
# Each compute job usually has data products that are required to be
# staged in to the execution site, materialized data products staged
# out to a final resting place, or staged to another job running at a
# different site. This property determines the underlying grid
# transfer tool that is used to manage the transfers.
#
# The * in the property name can be replaced to achieve finer grained
# control to dictate what type of transfer jobs need to be managed
# with which grid transfer tool.
#
# Usually,the arguments with which the client is invoked can be
# specified by
# <screen>
# - the property pegasus.transfer.arguments
# - associating the PEGASUS profile key transfer.arguments
# </screen>
#
#
# The table below illustrates all the possible variations of the
# property.
#
#
# <tb2>
# Property Name & Applies to </te>
# pegasus.transfer.stagein.impl & the stage in transfer jobs</te>
# pegasus.transfer.stageout.impl & the stage out transfer jobs</te>
# pegasus.transfer.inter.impl & the inter pool transfer jobs </te>
# pegasus.transfer.setup.impl & the setup transfer job</te>
# pegasus.transfer.*.impl & apply to types of transfer jobs </te>
# </tb2>
#
# Note: Since version 2.2.0 the worker package is staged automatically during
# staging of executables to the remote site. This is achieved
# by adding a setup transfer job to the workflow. The setup transfer job by
# default uses GUC to stage the data. The implementation to use can be
# configured by setting the property
# <screen>pegasus.transfer.setup.impl </screen>property.
# However, if you have pegasus.transfer.*.impl set in your properties file,
# then you need to set pegasus.transfer.setup.impl to GUC
#
#
# The various grid transfer tools that can be used to manage data
# transfers are explained below
#
# <variablelist>
# <varlistentry><term>Transfer</term>
# <listitem>
# <para>This results in pegasus-transfer to be used for transferring of files. It
# is a python based wrapper around various transfer clients like
# globus-url-copy, lcg-copy, wget, cp, ln . pegasus-transfer looks at
# source and destination url and figures out automatically which underlying
# client to use. pegasus-transfer is distributed with the PEGASUS and can
# be found at $PEGASUS_HOME/bin/pegasus-transfer.
#
# For remote sites, Pegasus constructs the default path to pegasus-transfer
# on the basis of PEGASUS_HOME env profile specified in the site catalog.
# To specify a different path to the pegasus-transfer client , users can
# add an entry into the transformation catalog with fully qualified logical
# name as pegasus::pegasus-transfer
# </para>
# </listitem></varlistentry>
# <varlistentry><term>GUC</term>
# <listitem>This refers to the new guc client that does multiple file
# transfers per invocation. The globus-url-copy client
# distributed with Globus 4.x is compatible with this mode.
# </listitem></varlistentry>
# </variablelist>
#
#
# pegasus.transfer.*.impl Transfer
# Property : pegasus.transfer.refiner
# System : Pegasus
# Type : enumeration
# Value[0] : Basic
# Value[1] : Cluster
# Default : Cluster
# Since : 2.0
# See also : pegasus.transfer.*.impl
#
# This property determines how the transfer nodes are added to the
# workflow. The various refiners differ in the how they link the
# various transfer jobs, and the number of transfer jobs that are
# created per compute jobs.
# <variablelist>
# <varlistentry><term>Basic</term>
# <listitem>This is a basic refinement strategy that adds a stage-in
# job per compute job and a stage-out per compute jobs. It is not
# recommended to use this , especially for large workflows where lots
# of stage-in jobs maybe created for a workflow. This is only recommended
# for experimental setups.
# </listitem></varlistentry>
# <varlistentry><term>Cluster</term>
# <listitem>
# <para>In this refinement strategy, clusters of stage-in and stageout jobs
# are created per level of the workflow. This workflow allows you to
# control the number of stagein and stageout jobs by associating pegasus
# profiles stagein.clusters and stageout.clusters with the jobs or in the
# site catalog for the staging sites.
# </para>
# </listitem></varlistentry>
# </variablelist>
#
#
# pegasus.transfer.refiner Cluster
# Property : pegasus.transfer.lite.*.impl
# System : Pegasus
# Type : enumeration
# Value[0] : Transfer
# Value[1] : Condor
# Default : Transfer
# Since : 4.4.0
# See also : pegasus.data.configuration
# See also : pegasus.execute.*.filesystem.local
#
# This property specifies the transfer tool to be used for
# staging of input and output data between the staging site and worker
# node filesystems in PegasusLite.
#
# Currently, the * in the property name CANNOT be replaced to achieve
# finer grained control to dictate what type of transfers need to
# be managed with which grid transfer tool.
#
#
# The various grid transfer tools that can be used to manage PegasusLite data
# transfers are explained below
#
# <variablelist>
# <varlistentry><term>Transfer</term>
# <listitem>
# <para>This results in pegasus-transfer to be used for transferring of files. It
# is a python based wrapper around various transfer clients like
# globus-url-copy, lcg-copy, wget, cp, ln . pegasus-transfer looks at
# source and destination url and figures out automatically which underlying
# client to use. pegasus-transfer is distributed with the PEGASUS and can
# be found at $PEGASUS_HOME/bin/pegasus-transfer.
#
# For remote sites, Pegasus constructs the default path to pegasus-transfer
# on the basis of PEGASUS_HOME env profile specified in the site catalog.
# To specify a different path to the pegasus-transfer client , users can
# add an entry into the transformation catalog with fully qualified logical
# name as pegasus::pegasus-transfer
# </para>
# </listitem></varlistentry>
# <varlistentry><term>Condor</term>
# <listitem>
# <para>This results in Condor file transfer mechanism to be used to transfer the
# input data files from the submit host directly to the worker node
# directories. This is used when running in pure Condor mode or in a Condor
# pool that does not have a shared filesystem between the nodes.
#
# When setting the PegasusLite transfers to Condor make sure that the
# following properties are also set
# <screen>
# pegasus.gridstart PegasusLite
# pegasus.execute.*.filesystem.local true
# </screen>
# Alternatively, you can set
# <screen>
# pegasus.data.configuration condorio
# </screen> in lieu of the above 3 properties.
#
# Also make sure that pegasus.gridstart is not set.
#
# Please refer to the section on "Condor Pool Without a Shared Filesystem"
# in the chapter on Planning and Submitting.
# </para>
# </listitem></varlistentry>
# </variablelist>
#
#
# pegasus.transfer.lite.*.impl Transfer
# Property : pegasus.transfer.arguments
# System : Pegasus
# Since : 2.0
# Type : String
# Default : (no default)
# See also : pegasus.transfer.lite.arguments
#
# This determines the extra arguments with which the transfer implementation is
# invoked. The transfer executable that is invoked is dependant upon
# the transfer mode that has been selected.
# The property can be overloaded by associated the pegasus profile key
# transfer.arguments either with the site in the site catalog or the
# corresponding transfer executable in the transformation catalog.
#
#
# pegasus.transfer.arguments (no default)
# Property : pegasus.transfer.threads
# System : Pegasus
# Since : 4.4
# Type : Integer
# Default : 2
# See also : pegasus.transfer.lite.threads
#
# This property set the number of threads pegasus-transfer uses to transfer
# the files. This property to applies to the separate data transfer nodes
# that are added by Pegasus to the executable workflow.
# The property can be overloaded by associated the pegasus profile key
# transfer.threads either with the site in the site catalog or the
# corresponding transfer executable in the transformation catalog.
#
#
# pegasus.transfer.threads 2
# Property : pegasus.transfer.lite.arguments
# System : Pegasus
# Since : 4.4
# Type : String
# Default : (no default)
# See also : pegasus.transfer.arguments
# See also : pegasus.transfer.lite.*.impl
#
# This determines the extra arguments with which the PegasusLite transfer
# implementation is invoked. The transfer executable that is invoked
# is dependant upon the PegasusLite transfer implementation that has
# been selected.
#
#
# pegasus.transfer.lite.arguments (no default)
# Property : pegasus.transfer.lite.threads
# System : Pegasus
# Since : 4.4
# Type : Integer
# Default : 1
# See also : pegasus.transfer.lite.arguments
#
# This property set the number of threads pegasus-transfer uses to transfer
# the files. This property applies to the pegasus-transfer invocations in
# the PegasusLite jobs in the nonsharedfs data configuration.
# The property can be overloaded by associated the pegasus profile key
# transfer.lite.threads either with the site in the site catalog or the
# DAX for the associate compute jobs.
#
#
# pegasus.transfer.lite.threads 1
# Property : pegasus.transfer.worker.package
# System : Pegasus
# Type : boolean
# Default : false
# Since : 3.0
# See also : pegasus.data.configuration
#
# By default, Pegasus relies on the worker package to be installed in a directory
# accessible to the worker nodes on the remote sites . Pegasus uses the value of
# PEGASUS_HOME environment profile in the site catalog for the remote sites, to then
# construct paths to pegasus auxillary executables like kickstart, pegasus-transfer,
# seqexec etc.
#
# If the Pegasus worker package is not installed on the remote sites
# users can set this property to true to get Pegasus to deploy worker package on the
# nodes.
#
#
# In the case of sharedfs setup, the worker package is deployed on the shared scratch
# directory for the workflow , that is accessible to all the compute nodes of the
# remote sites.
#
# When running in nonsharefs environments, the worker package is first brought to the
# submit directory and then transferred to the worker node filesystem using Condor
# file IO.
#
#
# pegasus.transfer.worker.package false
# Property : pegasus.transfer.links
# System : Pegasus
# Type : boolean
# Default : false
# Since : 2.0
# See also : pegasus.transfer
#
#
# If this is set, and the transfer implementation is set to Transfer
# i.e. using the transfer executable distributed with the PEGASUS.
# On setting this property, if Pegasus while fetching data from the
# Replica Catalog sees a pool attribute associated with the PFN that matches
# the execution pool on which the data has to be transferred to,
# Pegasus instead of the URL returned by the Replica Catalog replaces it with
# a file based URL. This is based on the assumption that the if the pools match the
# filesystems are visible to the remote execution directory where
# input data resides.
# On seeing both the source and destination urls as file based URLs
# the transfer executable spawns a job that creates a symbolic link
# by calling ln -s on the remote pool.
#
#
#
# pegasus.transfer.links false
# Property : pegasus.transfer.*.remote.sites
# System : Pegasus
# Type : comma separated list of sites
# Default : no default
# Since : 2.0
#
# By default Pegasus looks at the source and destination URL's for to determine
# whether the associated transfer job runs on the submit host or the head node
# of a remote site, with preference set to run a transfer job to run on submit
# host.
#
# Pegasus will run transfer jobs on the remote sites
#
# <screen>
# - if the file server for the compute site is a file server i.e url prefix file://
# - symlink jobs need to be added that require the symlink transfer jobs to
# be run remotely.
# </screen>
#
# This property can be used to change the default behaviour of Pegasus and force pegasus
# to run different types of transfer jobs for the sites specified on the remote site.
#
# The table below illustrates all the possible variations of the
# property.
#
#
# <tb2>
# Property Name & Applies to </te>
# pegasus.transfer.stagein.remote.sites & the stage in transfer jobs</te>
# pegasus.transfer.stageout.remote.sites & the stage out transfer jobs</te>
# pegasus.transfer.inter.remote.sites & the inter pool transfer jobs </te>
# pegasus.transfer.*.remote.sites & apply to types of transfer jobs </te>
# </tb2>
#
# In addition * can be specified as a property value, to designate
# that it applies to all sites.
#
# pegasus.transfer.*.remote.sites (no default)
# Property : pegasus.transfer.staging.delimiter
# System : Pegasus
# Since : 2.0
# Type : String
# Default : :
# See also : pegasus.transformation.selector
#
# Pegasus supports executable staging as part of the
# workflow. Currently staging of statically linked executables is
# supported only. An executable is normally staged to the work
# directory for the workflow/partition on the remote site. The
# basename of the staged executable is derived from the namespace,name
# and version of the transformation in the transformation
# catalog. This property sets the delimiter that is used for the
# construction of the name of the staged executable.
#
#
# pegasus.transfer.staging.delimiter :
# Property : pegasus.transfer.disable.chmod.sites
# System : Pegasus
# Since : 2.0
# Type : comma separated list of sites
# Default : no default
#
#
# During staging of executables to remote sites, chmod jobs are
# added to the workflow. These jobs run on the remote sites and do a
# chmod on the staged executable. For some sites, this maynot be
# required. The permissions might be preserved, or there maybe an
# automatic mechanism that does it.
#
# This property allows you to specify the list of sites, where you do
# not want the chmod jobs to be executed. For those sites, the chmod
# jobs are replaced by NoOP jobs. The NoOP jobs are executed by
# Condor, and instead will immediately have a terminate event written
# to the job log file and removed from the queue.
#
#
#
#
# pegasus.transfer.disable.chmod.sites (no default)
# Property : pegasus.transfer.setup.source.base.url
# System : Pegasus
# Type : URL
# Default : no default
# Since : 2.3
#
#
# This property specifies the base URL to the directory containing the
# Pegasus worker package builds. During Staging of Executable, the
# Pegasus Worker Package is also staged to the remote site. The worker
# packages are by default pulled from the http server at pegasus.isi.edu.
# This property can be used to override the location from where the worker
# package are staged. This maybe required if the remote computes sites don't
# allows files transfers from a http server.
#
# pegasus.transfer.setup.source.base.url (no default)
#
# SECTION "GRIDSTART AND EXITCODE PROPERTIES"
#
# Property : pegasus.gridstart
# System : Pegasus
# Since : 2.0
# Type : enumeration
# Value[0] : Kickstart
# Value[1] : None
# Value[2] : PegasusLite
# Default : Kickstart
# See also : pegasus.execute.*.filesystem.local
#
# Jobs that are launched on the grid maybe wrapped in a wrapper
# executable/script that enables information about about the
# execution, resource consumption, and - most importantly - the
# exitcode of the remote application.
# At present, a job scheduled on a remote site is launched with a
# gridstart if site catalog has the corresponding gridlaunch attribute
# set and the job being launched is not MPI.
#
# Users can explicitly decide what gridstart to use for a job, by
# associating the pegasus profile key named gridstart with the job.
#
# <variablelist>
# <varlistentry><term>Kickstart</term>
# <listitem>In this mode, all the jobs are lauched via kickstart. The
# kickstart executable is a light-weight program
# which connects the stdin,stdout and stderr filehandles for
# PEGASUS jobs on the remote site. Kickstart is an executable
# distributed with PEGASUS that can generally be found at
# ${pegasus.home.bin}/kickstart.
# </listitem></varlistentry>
# <varlistentry><term>None</term>
# <listitem>In this mode, all the jobs are launched directly on
# the remote site. Each job's stdin,stdout and stderr are
# connected to condor commands in a manner to ensure that they are
# sent back to the submit host.
# </listitem></varlistentry>
# <varlistentry><term>PegasusLite</term>
# <listitem>In this mode, the compute jobs are wrapped by PegasusLite instances.
# PegasusLite instance is a bash script, that is launced on the compute node.
# It determins at runtime the directory a job needs to execute in, pulls in data
# from the staging site , launches the job, pushes out the data and cleans up the
# directory after execution.
# </listitem></varlistentry>
# </variablelist>
#
#
#
# pegasus.gridstart Kickstart
# Property : pegasus.gridstart.kickstart.set.xbit
# System : Pegasus
# Since : 2.4
# Type : Boolean
# Default : false
# See also : pegasus.transfer.disable.chmod.sites
#
#
# Kickstart has an option to set the X bit on an executable before it
# launches it on the remote site. In case of staging of executables,
# by default chmod jobs are launched that set the x bit of the user
# executables staged to a remote site.
#
# On setting this property to true, kickstart gridstart module adds a
# -X option to kickstart arguments. The -X arguments tells kickstart
# to set the x bit of the executable before launching it.
#
# User should usually disable the chmod jobs by setting the property
# pegasus.transfer.disable.chmod.sites , if they set this property
# to true.
#
#
# pegasus.gridstart.kickstart.set.xbit false
# Property : pegasus.gridstart.kickstart.stat
# System : Pegasus
# Since : 2.1
# Type : Boolean
# Default : false
# See also : pegasus.gridstart.generate.lof
#
#
# Kickstart has an option to stat the input files and the output
# files. The stat information is collected in the XML record generated
# by kickstart. Since stat is an expensive operation, it is not turned
# on by on. Set this property to true if you want to see stat
# information for the input files and output files of a job in it's
# kickstart output.
#
#
#
#
# pegasus.gridstart.kickstart.stat false
# Property : pegasus.gridstart.generate.lof
# System : Pegasus
# Since : 2.1
# Type : Boolean
# Default : false
# See also : pegasus.gridstart.kickstart.stat
#
#
# For the stat option for kickstart, we generate 2 lof ( list of
# filenames ) files for each job. One lof file containing the input
# lfn's for the job, and the other containing output lfn's for the
# job.
# In some cases, it maybe beneficial to have these lof files generated
# but not do the actual stat. This property allows you to generate the
# lof files without triggering the stat in kickstart invocations.
#
#
#
#
# pegasus.gridstart.generate.lof false
# Property : pegasus.gridstart.invoke.always
# System : Pegasus
# Since : 2.0
# Type : Boolean
# Default : false
# See also : pegasus.gridstart.invoke.length
#
# Condor has a limit in it, that restricts the length of arguments to
# an executable to 4K. To get around this limit, you can trigger
# Kickstart to be invoked with the -I option. In this case, an
# arguments file is prepared per job that is transferred to the remote
# end via the Condor file transfer mechanism. This way the arguments
# to the executable are not specified in the condor submit file for
# the job. This property specifies whether you want to use the invoke
# option always for all jobs, or want it to be triggered only when the
# argument string is determined to be greater than a certain limit.
#
#
# pegasus.gridstart.invoke.always false
# Property : pegasus.gridstart.invoke.length
# System : Pegasus
# Since : 2.0
# Type : Long
# Default : 4000
# See also : pegasus.gridstart.invoke.always
#
# Gridstart is automatically invoked with the -I option, if it is
# determined that the length of the arguments to be specified is going
# to be greater than a certain limit. By default this limit is set to
# 4K. However, it can overriden by specifying this property.
#
#
# pegasus.gridstart.invoke.length 4000
#
# SECTION "INTERFACE TO Condor and Condor DAGMan"
#
#
# The Condor DAGMan facility is usually activate using the
# condor_submit_dag command. However, many shapes of workflows have the
# ability to either overburden the submit host, or overflow remote
# gatekeeper hosts. While DAGMan provides throttles, unfortunately these
# can only be supplied on the command-line. Thus,PEGASUS provides a
# versatile wrapper to invoke DAGMan, called pegasus-submit-dag. It can be
# configured from the command-line, from user- and system properties,
# and by defaults.
# Property : pegasus.condor.logs.symlink
# System : Condor
# Type : Boolean
# Default : false
# Since : 3.0
#
# Starting 4.2.1 release, this property defaults to false. Prior to that
# it defaulted to true.
#
# If this property is set to true, then Pegasus will have the Condor
# common log [dagname]-0.log in the submit file as a symlink to a
# location in /tmp . You want to set this to true when your workflow
# submit directory is on the shared filesystem . You don't want the
# common log to get written to a shared filesystem. If the user knows
# for sure that the workflow submit directory is not on the shared filesystem,
# then the value to this property should be false.
#
#
#
# pegasus.condor.logs.symlink false
# Property : pegasus.condor.arguments.quote
# System : Condor
# Type : Boolean
# Default : true
# Since : 2.0
# Old Name : pegasus.condor.arguments.quote
#
# This property determines whether to apply the new Condor quoting
# rules for quoting the argument string. The new argument quoting
# rules appeared in Condor 6.7.xx series. We have verified it for
# 6.7.19 version. If you are using an old condor at the submit host,
# set this property to false.
#
#
#
# pegasus.scheduler.condor.arguments.quote true
# Property : pegasus.dagman.notify
# System : DAGman wrapper
# Type : Case-insensitive enumeration
# Value[0] : Complete
# Value[1] : Error
# Value[2] : Never
# Default : Never
# Document : http://www.cs.wisc.edu/condor/manual/v6.9/condor_submit_dag.html
# Document : http://www.cs.wisc.edu/condor/manual/v6.9/condor_submit.html
#
# The pegasus.dagman.nofity property has been deprecated in favor of the
# Pegasus notification framework. Please see the reference manual for
# details on how to get workflow notifications. pegasus.dagman.nofity
# will be removed in the an upcoming version of Pegasus.
#
# pegasus.dagman.notify Never
# Property : pegasus.dagman.verbose
# System : DAGman wrapper
# Type : Boolean
# Value[0] : false
# Value[1] : true
# Default : false
# Document : http://www.cs.wisc.edu/condor/manual/v6.9/condor_submit_dag.html
#
# The pegasus-submit-dag wrapper processes properties to set DAGMan
# commandline arguments. If set and true, the argument activates
# verbose output in case of DAGMan errors.
#
# pegasus.dagman.verbose false
# Property : pegasus.dagman.[category].maxjobs
# System : DAGman wrapper
# Type : Integer
# Since : 2.2
# Default : no default
# Document : http://vtcpc.isi.edu/pegasus/index.php/ChangeLog\#Support_for_DAGMan_node_categories
#
# DAGMan now allows for the nodes in the DAG to be grouped in
# category. The tuning parameters like maxjobs then can be applied per
# category instead of being applied to the whole workflow. To use this
# facility users need to associate the dagman profile key named
# category with their jobs. The value of the key is the category to
# which the job belongs to.
#
# You can then use this property to specify the value for a
# category. For the above example you will set
# pegasus.dagman.short-running.maxjobs
#
# pegasus.dagman.[category].maxjobs no default
#
# SECTION "MONITORING PROPERTIES"
#
# Property : pegasus.monitord.events
# System : Pegasus-monitord
# Type : Boolean
# Default : true
# Since : 3.0.2
# See Also : pegasus.monitord.output
#
# This property determines whether pegasus-monitord generates log
# events. If log events are disabled using this property, no bp file,
# or database will be created, even if the pegasus.monitord.output
# property is specified.
#
#
#
# pegasus.monitord.events true
# Property : pegasus.monitord.output
# System : Pegasus-monitord
# Type : String
# Since : 3.0.2
# See Also : pegasus.monitord.events
#
# This property specifies the destination for generated log events in
# pegasus-monitord. By default, events are stored in a sqlite database
# in the workflow directory, which will be created with the workflow's
# name, and a ".stampede.db" extension. Users can specify an
# alternative database by using a SQLAlchemy connection
# string. Details are available at:
# <screen>
# http://www.sqlalchemy.org/docs/05/reference/dialects/index.html
# </screen>
# It is important to note that users will need to have the appropriate
# db interface library installed. Which is to say, SQLAlchemy is a
# wrapper around the mysql interface library (for instance), it does
# not provide a MySQL driver itself. The Pegasus distribution
# includes both SQLAlchemy and the SQLite Python driver.
# As a final note, it is important to mention that unlike when using
# SQLite databases, using SQLAlchemy with other database servers,
# e.g. MySQL or Postgres , the target database needs to exist.
# Users can also specify a file name using this property in order to
# create a file with the log events.
#
# Example values for the SQLAlchemy connection string for various end points
# are listed below
#
# <tb2>
# SQL Alchemy End Point & Example Value </te>
# Netlogger BP File & file:///submit/dir/myworkflow.bp</te>
# SQL Lite Database & sqlite:///submit/dir/myworkflow.db</te>
# MySQL Database & mysql://user:password@host:port/databasename</te>
# </tb2>
#
#
#
# pegasus.monitord.output (no default)
# Property : pegasus.dashboard.output
# System : Pegasus-monitord
# Type : String
# Since : 4.2
# See Also : pegasus.monitord.output
#
# This property specifies the destination for the workflow dashboard database.
# By default, the workflow dashboard datbase defaults to a sqlite database
# named workflow.db in the $HOME/.pegasus directory. This is database is shared
# for all workflows run as a particular user.
# Users can specify an alternative database by using a SQLAlchemy connection
# string. Details are available at:
# <screen>
# http://www.sqlalchemy.org/docs/05/reference/dialects/index.html
# </screen>
# It is important to note that users will need to have the appropriate
# db interface library installed. Which is to say, SQLAlchemy is a
# wrapper around the mysql interface library (for instance), it does
# not provide a MySQL driver itself. The Pegasus distribution
# includes both SQLAlchemy and the SQLite Python driver.
# As a final note, it is important to mention that unlike when using
# SQLite databases, using SQLAlchemy with other database servers,
# e.g. MySQL or Postgres , the target database needs to exist.
# Users can also specify a file name using this property in order to
# create a file with the log events.
#
# Example values for the SQLAlchemy connection string for various end points
# are listed below
#
# <tb2>
# SQL Alchemy End Point & Example Value </te>
# SQL Lite Database & sqlite:///shared/myworkflow.db</te>
# MySQL Database & mysql://user:password@host:port/databasename</te>
# </tb2>
#
#
#
# pegasus.dashboard.output (no default)
# Property : pegasus.monitord.notifications
# System : Pegasus-monitord
# Type : Boolean
# Default : true
# Since : 3.1
# See Also : pegasus.monitord.notifications.max
# See Also : pegasus.monitord.notifications.timeout
#
# This property determines whether pegasus-monitord processes
# notifications. When notifications are enabled, pegasus-monitord will
# parse the .notify file generated by pegasus-plan and will invoke
# notification scripts whenever conditions matches one of the
# notifications.
#
#
#
# pegasus.monitord.notifications true
# Property : pegasus.monitord.notifications.max
# System : Pegasus-monitord
# Type : Integer
# Default : 10
# Since : 3.1
# See Also : pegasus.monitord.notifications
# See Also : pegasus.monitord.notifications.timeout
#
# This property determines how many notification scripts
# pegasus-monitord will call concurrently. Upon reaching this limit,
# pegasus-monitord will wait for one notification script to finish
# before issuing another one. This is a way to keep the number of
# processes under control at the submit host. Setting this property to
# 0 will disable notifications completely.
#
#
#
# pegasus.monitord.notifications.max 10
# Property : pegasus.monitord.notifications.timeout
# System : Pegasus-monitord
# Type : Integer
# Default : 0
# Since : 3.1
# See Also : pegasus.monitord.notifications
# See Also : pegasus.monitord.notifications.max
#
# This property determines how long will pegasus-monitord let
# notification scripts run before terminating them. When this property
# is set to 0 (default), pegasus-monitord will not terminate any
# notification scripts, letting them run indefinitely. If some
# notification scripts missbehave, this has the potential problem of
# starving pegasus-monitord's notification slots (see the
# pegasus.monitord.notifications.max property), and block further
# notifications. In addition, users should be aware that
# pegasus-monitord will not exit until all notification scripts are
# finished.
#
#
#
# pegasus.monitord.notifications.timeout 0
# Property : pegasus.monitord.stdout.disable.parsing
# System : Pegasus-monitord
# Type : Boolean
# Default : False
# Since : 3.1.1
#
# By default, pegasus-monitord parses the stdout/stderr section of the
# kickstart to populate the applications captured stdout and stderr in
# the job instance table for the stampede schema. For large workflows,
# this may slow down monitord especially if the application is
# generating a lot of output to it's stdout and stderr. This property,
# can be used to turn of the database population.
#
#
#
#
# pegasus.monitord.stdout.disable.parsing false
#
# SECTION "JOB CLUSTERING PROPERTIES"
#
# Property : pegasus.clusterer.job.aggregator
# System : Job Clustering
# Since : 2.0
# Type : String
# Value[0] : seqexec
# Value[1] : mpiexec
# Default : seqexec
#
# A large number of workflows executed through the Virtual Data
# System, are composed of several jobs that run for only a few seconds
# or so. The overhead of running any job on the grid is usually 60
# seconds or more. Hence, it makes sense to collapse small independent
# jobs into a larger job.
# This property determines, the executable that will be used for
# running the larger job on the remote site.
#
# <variablelist>
# <varlistentry><term>seqexec</term>
# <listitem>In this mode, the executable used to run the merged job is
# seqexec that runs each of the smaller jobs sequentially on the
# same node. The executable "seqexec" is a PEGASUS tool distributed
# in the PEGASUS worker package, and can be usually found at
# {pegasus.home}/bin/seqexec.
# </listitem></varlistentry>
# <varlistentry><term>mpiexec</term>
# <listitem>In this mode, the executable used to run the merged job is
# mpiexec that runs the smaller jobs via mpi on n nodes where n
# is the nodecount associated with the merged job. The executable
# "mpiexec" is a PEGASUS tool distributed in the PEGASUS worker package,
# and can be usually found at {pegasus.home}/bin/mpiexec.
# </listitem></varlistentry>
# </variablelist>
#
#
#
# pegasus.clusterer.job.aggregator seqexec
# Property : pegasus.clusterer.job.aggregator.seqexec.log
# System : Job Clustering
# Type : Boolean
# Default : false
# Since : 2.3
# See also : pegasus.clusterer.job.aggregator
# See also : pegasus.clusterer.job.aggregator.seqexec.log.global
#
#
# Seqexec logs the progress of the jobs that are being run by it in a
# progress file on the remote cluster where it is executed.
#
# This property sets the Boolean flag, that indicates whether to turn
# on the logging or not.
#
#
# pegasus.clusterer.job.aggregator.seqexec.log false
# Property : pegasus.clusterer.job.aggregator.seqexec.log.global
# System : Job Clustering
# Type : Boolean
# Default : true
# Since : 2.3
# See also : pegasus.clusterer.job.aggregator
# See also : pegasus.clusterer.job.aggregator.seqexec.log
# Old Name : pegasus.clusterer.job.aggregator.seqexec.hasgloballog
#
#
# Seqexec logs the progress of the jobs that are being run by it in a
# progress file on the remote cluster where it is executed. The
# progress log is useful for you to track the progress of your
# computations and remote grid debugging. The progress log file can be
# shared by multiple seqexec jobs that are running on a particular
# cluster as part of the same workflow. Or it can be per job.
#
# This property sets the Boolean flag, that indicates whether to have
# a single global log for all the seqexec jobs on a particular cluster
# or progress log per job.
#
#
#
# pegasus.clusterer.job.aggregator.seqexec.log.global true
# Property : pegasus.clusterer.job.aggregator.seqexec.firstjobfail
# System : Job Clustering
# Type : Boolean
# Default : true
# Since : 2.2
# See also : pegasus.clusterer.job.aggregator
#
# By default seqexec does not stop execution even if one of the
# clustered jobs it is executing fails. This is because seqexec tries
# to get as much work done as possible.
#
# This property sets the Boolean flag, that indicates whether to make
# seqexec stop on the first job failure it detects.
#
#
#
# pegasus.clusterer.job.aggregator.seqexec.firstjobfail false
# Property : pegasus.clusterer.label.key
# System : Job Clustering
# Type : String
# Default : label
# Since : 2.0
# See also : pegasus.partitioner.label.key
#
# While clustering jobs in the workflow into larger jobs, you can
# optionally label your graph to control which jobs are clustered and
# to which clustered job they belong. This done using a label based
# clustering scheme and is done by associating a profile/label key in
# the PEGASUS namespace with the jobs in the DAX. Each job that has the
# same value/label value for this profile key, is put in the same
# clustered job.
#
# This property allows you to specify the PEGASUS profile key that you
# want to use for label based clustering.
#
#
# pegasus.clusterer.label.key label
#
# SECTION "LOGGING PROPERTIES"
#
# Property : pegasus.log.manager
# System : Pegasus
# Since : 2.2.0
# Type : Enumeration
# Value[0] : Default
# Value[1] : Log4j
# Default : Default
# See also : pegasus.log.manager.formatter
#
#
# This property sets the logging implementation to use for logging.
#
# <variablelist>
# <varlistentry><term>Default</term>
# <listitem>This implementation refers to the legacy Pegasus logger, that
# logs directly to stdout and stderr. It however, does have the
# concept of levels similar to log4j or syslog.
# </listitem></varlistentry>
# <varlistentry><term>Log4j</term>
# <listitem>This implementation, uses Log4j to log messages. The log4j
# properties can be specified in a properties file, the location of
# which is specified by the property
# <screen>
# pegasus.log.manager.log4j.conf
# </screen>
# </listitem></varlistentry>
# </variablelist>
#
#
# pegasus.log.manager Default
# Property : pegasus.log.manager.formatter
# System : Pegasus
# Since : 2.2.0
# Type : Enumeration
# Value[0] : Simple
# Value[1] : Netlogger
# Default : Simple
# See also : pegasus.log.manager.formatter
#
#
# This property sets the formatter to use for formatting the log messages
# while logging.
#
# <variablelist>
# <varlistentry><term>Simple</term>
# <listitem>This formats the messages in a simple format. The messages are logged as
# is with minimal formatting. Below are sample log messages in this format
# while ranking a dax according to performance.
# <screen>
# event.pegasus.ranking dax.id se18-gda.dax - STARTED
# event.pegasus.parsing.dax dax.id se18-gda-nested.dax - STARTED
# event.pegasus.parsing.dax dax.id se18-gda-nested.dax - FINISHED
# job.id jobGDA
# job.id jobGDA query.name getpredicted performace time 10.00
# event.pegasus.ranking dax.id se18-gda.dax - FINISHED
# </screen>
# </listitem></varlistentry>
# <varlistentry><term>Netlogger</term>
# <listitem>
# <para>This formats the messages in the Netlogger format , that is based on key
# value pairs. The netlogger format is useful for loading the logs into a
# database to do some meaningful analysis. Below are sample log messages
# in this format while ranking a dax according to performance.
# <screen>
# ts=2008-09-06T12:26:20.100502Z event=event.pegasus.ranking.start \
# msgid=6bc49c1f-112e-4cdb-af54-3e0afb5d593c \
# eventId=event.pegasus.ranking_8d7c0a3c-9271-4c9c-a0f2-1fb57c6394d5 \
# dax.id=se18-gda.dax prog=Pegasus
#
# ts=2008-09-06T12:26:20.100750Z event=event.pegasus.parsing.dax.start \
# msgid=fed3ebdf-68e6-4711-8224-a16bb1ad2969 \
# eventId=event.pegasus.parsing.dax_887134a8-39cb-40f1-b11c-b49def0c5232\
# dax.id=se18-gda-nested.dax prog=Pegasus
#
# ts=2008-09-06T12:26:20.100894Z event=event.pegasus.parsing.dax.end \
# msgid=a81e92ba-27df-451f-bb2b-b60d232ed1ad \
# eventId=event.pegasus.parsing.dax_887134a8-39cb-40f1-b11c-b49def0c5232
#
# ts=2008-09-06T12:26:20.100395Z event=event.pegasus.ranking \
# msgid=4dcecb68-74fe-4fd5-aa9e-ea1cee88727d \
# eventId=event.pegasus.ranking_8d7c0a3c-9271-4c9c-a0f2-1fb57c6394d5 \
# job.id="jobGDA"
#
# ts=2008-09-06T12:26:20.100395Z event=event.pegasus.ranking \
# msgid=4dcecb68-74fe-4fd5-aa9e-ea1cee88727d \
# eventId=event.pegasus.ranking_8d7c0a3c-9271-4c9c-a0f2-1fb57c6394d5 \
# job.id="jobGDA" query.name="getpredicted performace" time="10.00"
#
# ts=2008-09-06T12:26:20.102003Z event=event.pegasus.ranking.end \
# msgid=31f50f39-efe2-47fc-9f4c-07121280cd64 \
# eventId=event.pegasus.ranking_8d7c0a3c-9271-4c9c-a0f2-1fb57c6394d5
# </screen>
# </para>
# </listitem></varlistentry>
# </variablelist>
#
#
# pegasus.log.manager.formatter Simple
# Property : pegasus.log.*
# System : Pegasus
# Since : 2.0
# Type : String
# Default : No default
#
#
# This property sets the path to the file where all the logging for
# Pegasus can be redirected to. Both stdout and stderr are logged to
# the file specified.
#
#
# pegasus.log.* no default
# Property : pegasus.log.memory.usage
# System : Pegasus
# Since : 4.3.4
# Type : Boolean
# Default : false
#
#
# This property if set to true, will result in the planner writing out
# JVM heap memory statistics at the end of the planning process at the
# INFO level. This is useful, if users want to fine tune their java
# memory settings by setting JAVA_HEAPMAX and JAVA_HEAPMIN for large
# workflows.
#
# pegasus.log.memory.usage false
# Property : pegasus.log.metrics
# System : Pegasus
# Since : 2.1.0
# Type : Boolean
# Default : true
# See also : pegasus.log.metrics.file
#
#
# This property enables the logging of certain planning and workflow
# metrics to a global log file. By default the file to which the
# metrics are logged is ${pegasus.home}/var/pegasus.log.
#
#
#
# pegasus.log.metrics true
# Property : pegasus.log.metrics.file
# System : Pegasus
# Since : 2.1.0
# Type : Boolean
# Default : ${pegasus.home}/var/pegasus.log
# See also : pegasus.log.metrics
#
#
# This property determines the file to which the workflow and planning
# metrics are logged if enabled.
#
#
#
# pegasus.log.metrics.file ${pegasus.home}/var/pegasus.log
# Property : pegasus.metrics.app
# System : Pegasus
# Since : 4.3.0
# Type : String
# See also : pegasus.log.metrics
#
#
# This property namespace allows users to pass application level metrics
# to the metrics server. The value of this property is the name of the
# application.
#
# Additional application specific attributes can be passed by using the
# prefix pegasus.metrics.app
# <screen>
# pegasus.metrics.app.[arribute-name] attribute-value
# </screen>
#
# Note: the attribute cannot be named name. This attribute is automatically
# assigned the value from pegasus.metrics.app
#
# pegasus.metrics.app (no default)
#
# SECTION "MISCELLANEOUS PROPERTIES"
#
# Property : pegasus.code.generator
# System : Pegasus
# Since : 3.0
# Type : enumeration
# Value[0] : Condor
# Value[1] : Shell
# Value[2] : PMC
# Default : Condor
#
# This property is used to load the appropriate Code Generator to use for
# writing out the executable workflow.
#
# <variablelist>
# <varlistentry><term>Condor</term>
# <listitem>
# This is the default code generator for Pegasus . This generator generates
# the executable workflow as a Condor DAG file and associated job submit files.
# The Condor DAG file is passed as input to Condor DAGMan for job execution.
# </listitem></varlistentry>
# <varlistentry><term>Shell</term>
# <listitem>
# This Code Generator generates the executable workflow as a shell script that
# can be executed on the submit host. While using this code generator, all the
# jobs should be mapped to site local i.e specify --sites local to pegasus-plan.
# </listitem></varlistentry>
# <varlistentry><term>PMC</term>
# <listitem>
# This Code Generator generates the executable workflow as a PMC task workflow.
# This is useful to run on platforms where it not feasible to run Condor such
# as the new XSEDE machines such as Blue Waters.
# In this mode, Pegasus will generate the executable workflow as a PMC task
# workflow and a sample PBS submit script that submits this workflow.
# </listitem></varlistentry>
# </variablelist>
#
#
# pegasus.code.generator Condor
# Property : pegasus.register
# System : Pegasus
# Since : 4.1.0
# Type : Boolean
# Default : true
#
# Pegasus creates registration jobs to register the output files in the replica
# catalog. An output file is registered only if
#
# 1) a user has configured a replica catalog in the properties
# 2) the register flags for the output files in the DAX are set to true
#
#
# This property can be used to turn off the creation of the registration jobs
# even though the files maybe marked to be registered in the replica catalog.
#
#
# pegasus.register true
# Property : pegasus.job.priority.assign
# System : Pegasus
# Since : 3.0.3
# Type : Boolean
# Default : true
#
# This property can be used to turn off the default level based condor priorities
# that are assigned to jobs in the executable workflow.
#
#
# pegasus.job.priority.assign true
# Property : pegasus.file.cleanup.strategy
# System : Pegasus
# Since : 2.2
# Type : enumeration
# Value[0] : InPlace
# Default : InPlace
#
# This property is used to select the strategy of how the the cleanup
# nodes are added to the executable workflow.
#
#
# <variablelist>
# <varlistentry><term>InPlace</term>
# <listitem>
# This is the only mode available .
# </listitem></varlistentry>
# </variablelist>
#
#
# pegasus.file.cleanup.strategy InPlace
# Property : pegasus.file.cleanup.impl
# System : Pegasus
# Since : 2.2
# Type : enumeration
# Value[0] : Cleanup
# Value[1] : RM
# Value[2] : S3
# Default : Cleanup
#
# This property is used to select the executable that is used to
# create the working directory on the compute sites.
#
#
# <variablelist>
# <varlistentry><term>Cleanup</term>
# <listitem>
# The default executable that is used to delete files is the
# dirmanager executable shipped with Pegasus. It is found at
# $PEGASUS_HOME/bin/dirmanager in the pegasus distribution.
# An entry for transformation pegasus::dirmanager needs
# to exist in the Transformation Catalog or the PEGASUS_HOME
# environment variable should be specified in the site catalog for
# the sites for this mode to work.
# </listitem></varlistentry>
# <varlistentry><term>RM</term>
# <listitem>
# This mode results in the rm executable to be used to delete files
# from remote directories. The rm executable is standard on *nix
# systems and is usually found at /bin/rm location.
# </listitem></varlistentry>
# <varlistentry><term>S3</term>
# <listitem>
# This mode is used to delete files/objects from the buckets in S3
# instead of a directory. This should be set when running workflows
# on Amazon EC2. This implementation relies on s3cmd command line
# client to create the bucket. An entry for transformation
# amazon::s3cmd needs to exist in the Transformation Catalog for
# this to work.
# </listitem></varlistentry>
# </variablelist>
#
#
# pegasus.file.cleanup.impl Cleanup
# Property : pegasus.file.cleanup.clusters.num
# System : Pegasus
# Since : 4.2
# Type : Integer
# Default : 2
#
# In case of the InPlace strategy for adding the cleanup nodes to the
# workflow, this property specifies the maximum number of cleanup
# jobs that are added to the executable workflow on each level.
#
#
# pegasus.file.cleanup.clusters.num 2
# Property : pegasus.file.cleanup.clusters.size
# System : Pegasus
# Since : 4.2.1
# Type : Integer
# Default : 2
#
# In case of the InPlace strategy this property sets the number of
# cleanup jobs that get clustered into a bigger cleanup job.
# This parameters is only used if pegasus.file.cleanup.clusters.num
# is not set.
#
# pegasus.file.cleanup.clusters.size ( no default )
# Property : pegasus.file.cleanup.scope
# System : Pegasus
# Since : 2.3.0
# Type : enumeration
# Value[0] : fullahead
# Value[1] : deferred
# Default : fullahead
#
# By default in case of deferred planning InPlace file cleanup is turned OFF.
# This is because the cleanup algorithm does not work across partitions.
# This property can be used to turn on the cleanup in case of deferred planning.
# <variablelist>
# <varlistentry><term>fullahead</term>
# <listitem>
# This is the default scope. The pegasus cleanup algorithm does not work
# across partitions in deferred planning. Hence the cleanup is always turned
# OFF , when deferred planning occurs and cleanup scope is set to full ahead.
# </listitem></varlistentry>
# <varlistentry><term>deferred</term>
# <listitem>
# If the scope is set to deferred, then Pegasus will not disable file cleanup
# in case of deferred planning. This is useful for scenarios where the
# partitions themselves are independant ( i.e. dont share files ). Even if
# the scope is set to deferred, users can turn off cleanup by specifying
# --nocleanup option to pegasus-plan.
# </listitem></varlistentry>
# </variablelist>
#
#
#
# pegasus.file.cleanup.scope fullahead
# Property : pegasus.catalog.transformation.mapper
# System : Staging of Executables
# Since : 2.0
# Type : enumeration
# Value[0] : All
# Value[1] : Installed
# Value[2] : Staged
# Value[3] : Submit
# Default : All
# See also : pegasus.transformation.selector
#
#
# Pegasus now supports transfer of statically linked executables as
# part of the concrete workflow. At present, there is only support for
# staging of executables referred to by the compute jobs specified in
# the DAX file.
# Pegasus determines the source locations of the binaries from the
# transformation catalog, where it searches for entries of type
# STATIC_BINARY for a particular architecture type. The PFN for these
# entries should refer to a globus-url-copy valid and accessible
# remote URL.
# For transfer of executables, Pegasus constructs a soft state map
# that resides on top of the transformation catalog, that helps in
# determining the locations from where an executable can be staged to
# the remote site.
#
# This property determines, how that map is created.
# <variablelist>
# <varlistentry><term>All</term>
# <listitem>In this mode, all sources with entries of type STATIC_BINARY
# for a particular transformation are considered valid sources for
# the transfer of executables. This the most general mode, and
# results in the constructing the map as a result of the cartesian
# product of the matches.
# </listitem></varlistentry>
# <varlistentry><term>Installed</term>
# <listitem>In this mode, only entries that are of type INSTALLED
# are used while constructing the soft state map. This results in
# Pegasus never doing any transfer of executables as part of the
# workflow. It always prefers the installed executables at the remote
# sites.
# </listitem></varlistentry>
# <varlistentry><term>Staged</term>
# <listitem>In this mode, only entries that are of type STATIC_BINARY
# are used while constructing the soft state map. This results in
# the concrete workflow referring only to the staged executables,
# irrespective of the fact that the executables are already
# installed at the remote end.
# </listitem></varlistentry>
# <varlistentry><term>Submit</term>
# <listitem>In this mode, only entries that are of type STATIC_BINARY
# and reside at the submit host (pool local), are used while
# constructing the soft state map. This is especially helpful,
# when the user wants to use the latest compute code for his
# computations on the grid and that relies on his submit
# host.
# </listitem></varlistentry>
# </variablelist>
#
# pegasus.catalog.transformation.mapper All
# Property : pegasus.selector.transformation
# System : Staging of Executables
# Since : 2.0
# Type : enumeration
# Value[0] : Random
# Value[1] : Installed
# Value[2] : Staged
# Value[3] : Submit
# Default : Random
# See also : pegasus.catalog.transformation
#
#
# In case of transfer of executables, Pegasus could have various
# transformations to select from when it schedules to run a particular
# compute job at a remote site. For e.g it can have the choice of
# staging an executable from a particular remote pool, from the local
# (submit host) only, use the one that is installed on the remote site
# only.
#
# This property determines, how a transformation amongst the various
# candidate transformations is selected, and is applied after the
# property pegasus.tc has been applied. For e.g specifying
# pegasus.tc as Staged and then pegasus.transformation.selector as
# INSTALLED does not work, as by the time this property is applied,
# the soft state map only has entries of type STAGED.
#
# <variablelist>
# <varlistentry><term>Random</term>
# <listitem>In this mode, a random matching candidate transformation
# is selected to be staged to the remote execution pool.
# </listitem></varlistentry>
# <varlistentry><term>Installed</term>
# <listitem>In this mode, only entries that are of type INSTALLED
# are selected. This means that the concrete workflow only refers
# to the transformations already pre installed on the remote
# pools.
# </listitem></varlistentry>
# <varlistentry><term>Staged</term>
# <listitem>In this mode, only entries that are of type STATIC_BINARY
# are selected, ignoring the ones that are installed at the remote
# site.
# </listitem></varlistentry>
# <varlistentry><term>Submit</term>
# <listitem>In this mode, only entries that are of type STATIC_BINARY
# and reside at the submit host (pool local), are selected as
# sources for staging the executables to the remote execution
# pools.
# </listitem></varlistentry>
# </variablelist>
#
# pegasus.selector.transformation Random
# Property : pegasus.execute.*.filesystem.local
# System : Pegasus
# Type : Boolean
# Default : false
# Since : 2.1.0
# See also : pegasus.data.configuration
#
#
# Normally, Pegasus transfers the data to and from a directory on the
# shared filesystem on the head node of a compute site. The directory
# needs to be visible to both the head node and the worker nodes for
# the compute jobs to execute correctly.
#
# By setting this property to true, you can get Pegasus to execute jobs
# on the worker node filesystem. In this case, when the jobs are
# launched on the worker nodes, the jobs grab the input data from
# the workflow specific execution directory on the compute site and
# push the output data to the same directory after completion.
# The transfer of data to and from the worker node directory is referred
# to as PegasusLite Data Transfers.
#
#
#
#
# pegasus.execute.*.filesystem.local false
# Property : pegasus.parser.dax.preserver.linebreaks
# System : Pegasus
# Type : Boolean
# Default : false
# Since : 2.2.0
#
#
# The DAX Parser normally does not preserve line breaks while parsing the
# CDATA section that appears in the arguments section of the job element
# in the DAX. On setting this to true, the DAX Parser preserves any line
# line breaks that appear in the CDATA section.
#
# pegasus.parser.dax.preserver.linebreaks false
# Property : pegasus.parser.dax.data.dependencies
# System : Pegasus
# Type : Boolean
# Default : true
# Since : 4.4.0
#
#
# If this property is set to true, then the planner will automatically
# add edges between jobs in the DAX on the basis of exisitng data
# dependencies between jobs. For example, if a JobA generates an output
# file that is listed as input for JobB, then the planner will
# automatically add an edge between JobA and JobB.
#
# pegasus.parser.dax.data.dependencies true
|