-
Notifications
You must be signed in to change notification settings - Fork 1
/
config.yaml
112 lines (98 loc) · 4.54 KB
/
config.yaml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
# Options for display and rendering
user_interface:
# Call out to the `stty` program to determine terminal size, instead of
# relying on what is reported by the curses library. In some cases,
# the curses library fails to update on SIGWINCH signals. If the
# `plotman interactive` curses interface does not properly adjust when
# you resize the terminal window, you can try setting this to True.
use_stty_size: True
# Where to plot and log.
directories:
# One directory in which to store all plot job logs (the STDOUT/
# STDERR of all plot jobs). In order to monitor progress, plotman
# reads these logs on a regular basis, so using a fast drive is
# recommended.
log: /farm/logs
# One or more directories to use as tmp dirs for plotting. The
# scheduler will use all of them and distribute jobs among them.
# It assumes that IO is independent for each one (i.e., that each
# one is on a different physical device).
#
# If multiple directories share a common prefix, reports will
# abbreviate and show just the uniquely identifying suffix.
tmp:
# - /farm/yards/900
# - /farm/yards/901
# Optional: Allows overriding some characteristics of certain tmp
# directories. This contains a map of tmp directory names to
# attributes. If a tmp directory and attribute is not listed here,
# it uses the default attribute setting from the main configuration.
#
# Currently support override parameters:
# - tmpdir_max_jobs
#tmp_overrides:
# In this example, /mnt/tmp/00 is larger than the other tmp
# dirs and it can hold more plots than the default.
#"/farm/yards/00":
#tmpdir_max_jobs: 5
# Optional: tmp2 directory. If specified, will be passed to
# chia plots create as -2. Only one tmp2 directory is supported.
# tmp2: /mnt/tmp/a
# One or more directories; the scheduler will use all of them.
# These again are presumed to be on independent physical devices,
# so writes (plot jobs) and reads (archivals) can be scheduled
# to minimize IO contention.
dst:
# - /farm/wagons/800
# - /farm/wagons/801
# - /farm/wagons/802
# - /farm/wagons/803
# Archival configuration. Optional; if you do not wish to run the
# archiving operation, comment this section out.
#
# Currently archival depends on an rsync daemon running on the remote
# host, and that the module is configured to match the local path.
# See code for details.
archive:
rsyncd_module: sites
rsyncd_path: /farm/sites
rsyncd_bwlimit: 80000 # Bandwidth limit in KB/s
rsyncd_host: chia
rsyncd_user: chia
# index: 0
# index: 1
# Plotting scheduling parameters
scheduling:
# tmpdir_stagger_phase_limit: 1
# tmpdir_stagger_phase_limit: 2
# tmpdir_stagger_phase_limit: 3
# tmpdir_stagger_phase_limit: 4
# Don't run a job on a particular temp dir until all existing jobs
# have progresed at least this far. Phase major corresponds to the
# plot phase, phase minor corresponds to the table or table pair
# in sequence.
tmpdir_stagger_phase_major: 2
tmpdir_stagger_phase_minor: 1
# Don't run more than this many jobs at a time on a single temp dir.
# tmpdir_max_jobs: 1
# tmpdir_max_jobs: 3
# tmpdir_max_jobs: 6
# tmpdir_max_jobs: 7
# tmpdir_max_jobs: 8
# tmpdir_max_jobs: 11
# Don't run any jobs (across all temp dirs) more often than this.
global_stagger_m: 90
# How often the daemon wakes to consider starting a new plot job
polling_time_s: 20
# Plotting parameters. These are pass-through parameters to chia plots create.
# See documentation at
# https://github.com/Chia-Network/chia-blockchain/wiki/CLI-Commands-Reference#create
plotting:
k: 32
e: False # Use -e plotting option
# n_threads: 1 # Threads per job
# n_threads: 2 # Threads per job
# n_threads: 3 # Threads per job
# n_threads: 4 # Threads per job
n_buckets: 128 # Number of buckets to split data into
job_buffer: 6000 # Per job memory