forked from lmacken/plow
-
Notifications
You must be signed in to change notification settings - Fork 0
/
plow.py
202 lines (166 loc) · 6.35 KB
/
plow.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
#!/usr/bin/env python3
"""
The Plow.
An efficient Chia plot mover.
Author: Luke Macken <[email protected]>
SPDX-License-Identifier: GPL-3.0-or-later
"""
import sys
import shutil
import random
import urllib.request
import asyncio
import aionotify
from pathlib import Path
from datetime import datetime
from collections import defaultdict
# Local plot sources
SOURCES = []
# Rsync destinations
# Examples: ["/mnt/HDD1", "192.168.1.10::hdd1"]
DESTS = []
# Shuffle plot destinations. Useful when using many plotters to decrease the odds
# of them copying to the same drive simultaneously.
SHUFFLE = True
# Rsync bandwidth limiting
BWLIMIT = None
# Optionally set the I/O scheduling class and priority
IONICE = None # "-c3" for "idle"
# Only send 1 plot at a time, regardless of source/dest.
ONE_AT_A_TIME = False
# Each plot source can have a lock, so we don't send more than one file from
# that origin at any given time.
ONE_PER_DRIVE = True
# Short & long sleep durations upon various error conditions
SLEEP_FOR = 60 * 3
SLEEP_FOR_LONG = 60 * 20
RSYNC_CMD = "rsync"
if SHUFFLE:
random.shuffle(DESTS)
if BWLIMIT:
RSYNC_FLAGS = f"--remove-source-files --preallocate --whole-file --bwlimit={BWLIMIT}"
else:
RSYNC_FLAGS = "--remove-source-files --preallocate --whole-file"
if IONICE:
RSYNC_CMD = "ionice {IONICE} rsync"
LOCK = asyncio.Lock() # Global ONE_AT_A_TIME lock
SRC_LOCKS = defaultdict(asyncio.Lock) # ONE_PER_DRIVE locks
async def plotfinder(paths, plot_queue, loop):
for path in paths:
for plot in Path(path).glob("**/*.plot"):
await plot_queue.put(plot)
await plotwatcher(paths, plot_queue, loop)
async def plotwatcher(paths, plot_queue, loop):
watcher = aionotify.Watcher()
for path in paths:
if not Path(path).exists():
print(f'! Path does not exist: {path}')
continue
print('watching', path)
watcher.watch(
alias=path,
path=path,
flags=aionotify.Flags.MOVED_TO,
)
await watcher.setup(loop)
while True:
event = await watcher.get_event()
print(event)
if event.name.endswith(".plot"):
plot_path = Path(event.alias) / event.name
await plot_queue.put(plot_path)
async def plow(dest, plot_queue, loop):
print(f"🧑🌾 plowing to {dest}")
while True:
try:
plot = await plot_queue.get()
cmd = f"{RSYNC_CMD} {RSYNC_FLAGS} {plot} {dest}"
# For local copies, we can check if there is enough space.
dest_path = Path(dest)
if dest_path.exists():
# Make sure it's actually a mount, and not our root filesystem.
if not dest_path.is_mount():
print(f"Farm destination {dest_path} is not mounted. Trying again later.")
await plot_queue.put(plot)
await asyncio.sleep(SLEEP_FOR)
continue
plot_size = plot.stat().st_size
dest_free = shutil.disk_usage(dest).free
if dest_free < plot_size:
print(f"Farm {dest} is full")
await plot_queue.put(plot)
# Just quit the worker entirely for this destination.
break
# One at a time, system-wide lock
if ONE_AT_A_TIME:
await LOCK.acquire()
# Only send one plot from each SSD at a time
if ONE_PER_DRIVE:
await SRC_LOCKS[plot.parent].acquire()
try:
print(f"🚜 {plot} ➡️ {dest}")
# Send a quick test copy to make sure we can write, or fail early.
test_cmd = f"rsync /etc/hostname {dest}"
proc = await asyncio.create_subprocess_shell(
test_cmd, stdout=asyncio.subprocess.PIPE, stderr=asyncio.subprocess.PIPE
)
stdout, stderr = await proc.communicate()
if proc.returncode != 0:
print(f"⁉️ {test_cmd!r} exited with {proc.returncode}")
await plot_queue.put(plot)
break
# Now rsync the real plot
proc = await asyncio.create_subprocess_shell(
cmd, stdout=asyncio.subprocess.PIPE, stderr=asyncio.subprocess.PIPE
)
start = datetime.now()
stdout, stderr = await proc.communicate()
finish = datetime.now()
finally:
if ONE_PER_DRIVE:
SRC_LOCKS[plot.parent].release()
if ONE_AT_A_TIME:
LOCK.release()
if proc.returncode == 0:
print(f"🏁 {cmd} ({finish - start})")
elif proc.returncode == 10: # Error in socket I/O
# Retry later.
print(f"⁉️ {cmd!r} exited with {proc.returncode} (error in socket I/O)")
await plot_queue.put(plot)
await asyncio.sleep(SLEEP_FOR_LONG)
elif proc.returncode in (11, 23): # Error in file I/O
# Most likely a full drive.
print(f"⁉️ {cmd!r} exited with {proc.returncode} (error in file I/O)")
await plot_queue.put(plot)
print(f"{dest} plow exiting")
break
else:
print(f"⁉️ {cmd!r} exited with {proc.returncode}")
await asyncio.sleep(SLEEP_FOR)
await plot_queue.put(plot)
print(f"{dest} plow exiting")
break
if stdout:
output = stdout.decode().strip()
if output:
print(f"{stdout.decode()}")
if stderr:
print(f"⁉️ {stderr.decode()}")
except Exception as e:
print(f"! {e}")
async def main(paths, loop):
plot_queue = asyncio.Queue()
futures = []
# Add plots to queue
futures.append(plotfinder(paths, plot_queue, loop))
# Fire up a worker for each plow
for dest in DESTS:
futures.append(plow(dest, plot_queue, loop))
print('🌱 Plow running...')
await asyncio.gather(*futures)
if __name__ == "__main__":
loop = asyncio.get_event_loop()
try:
loop.run_until_complete(main(SOURCES, loop))
except KeyboardInterrupt:
pass