forked from sqlpage/SQLPage
-
Notifications
You must be signed in to change notification settings - Fork 0
/
build.rs
138 lines (128 loc) · 4.78 KB
/
build.rs
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
use actix_rt::spawn;
use libflate::gzip;
use std::collections::hash_map::DefaultHasher;
use std::fs::File;
use std::hash::Hasher;
use std::io::Read;
use std::io::{BufRead, BufReader, Write};
use std::path::{Path, PathBuf};
use std::rc::Rc;
use std::time::Duration;
#[actix_rt::main]
async fn main() {
println!("cargo:rerun-if-changed=build.rs");
let c = Rc::new(make_client());
for h in [
spawn(download_deps(c.clone(), "sqlpage.js")),
spawn(download_deps(c.clone(), "sqlpage.css")),
spawn(download_deps(c.clone(), "tabler-icons.svg")),
spawn(download_deps(c.clone(), "apexcharts.js")),
spawn(download_deps(c.clone(), "tomselect.js")),
] {
h.await.unwrap();
}
}
fn make_client() -> awc::Client {
awc::ClientBuilder::new()
.timeout(Duration::from_secs(10))
.no_default_headers()
.finish()
}
/// Creates a file with inlined remote files included
async fn download_deps(client: Rc<awc::Client>, filename: &str) {
let path_in = format!("sqlpage/{}", filename);
let out_dir = PathBuf::from(std::env::var("OUT_DIR").unwrap());
let path_out: PathBuf = out_dir.join(filename);
// Generate outfile by reading infile and interpreting all comments
// like "/* !include https://... */" as a request to include the contents of
// the URL in the generated file.
println!("cargo:rerun-if-changed={}", path_in);
let original = File::open(path_in).unwrap();
process_input_file(&client, &path_out, original).await;
std::fs::write(
format!("{}.filename.txt", path_out.display()),
hashed_filename(&path_out),
)
.unwrap();
}
async fn process_input_file(client: &awc::Client, path_out: &Path, original: File) {
let mut outfile = gzip::Encoder::new(File::create(path_out).unwrap()).unwrap();
for l in BufReader::new(original).lines() {
let line = l.unwrap();
if line.starts_with("/* !include https://") {
let url = line
.trim_start_matches("/* !include ")
.trim_end_matches(" */");
copy_url_to_opened_file(client, url, &mut outfile).await;
outfile.write_all(b"\n").unwrap();
} else {
writeln!(outfile, "{}", line).unwrap();
}
}
outfile
.finish()
.as_result()
.expect("Unable to write compressed frontend asset");
}
async fn copy_url_to_opened_file(
client: &awc::Client,
url: &str,
outfile: &mut impl std::io::Write,
) {
// If the file has been downloaded manually, use it
let cached_file_path = make_url_path(url);
if !cached_file_path.exists() {
println!("cargo:warning=Downloading {url} to cache file {cached_file_path:?}.");
download_url_to_path(client, url, &cached_file_path).await;
println!("cargo:rerun-if-changed={}", cached_file_path.display());
}
copy_cached_to_opened_file(&cached_file_path, outfile);
}
fn copy_cached_to_opened_file(source: &Path, outfile: &mut impl std::io::Write) {
let reader = std::fs::File::open(source).unwrap();
let mut buf = std::io::BufReader::new(reader);
// Not async, but performance should not really matter here
std::io::copy(&mut buf, outfile).unwrap();
}
async fn download_url_to_path(client: &awc::Client, url: &str, path: &Path) {
let mut resp = client.get(url).send().await.unwrap_or_else(|err| {
let path = make_url_path(url);
panic!(
"We need to download external frontend dependencies to build the static frontend. \
Could not download {url}. You can manually download the file and place it in {path:?}\
{err}"
)
});
if resp.status() != 200 {
panic!("Received {} status code from {}", resp.status(), url);
}
let bytes = resp.body().limit(128 * 1024 * 1024).await.unwrap();
std::fs::write(path, &bytes)
.expect("Failed to write external frontend dependency to local file");
}
// Given a filename, creates a new unique filename based on the file contents
fn hashed_filename(path: &Path) -> String {
let mut file = File::open(path).unwrap();
let mut buf = [0u8; 4096];
let mut hasher = DefaultHasher::new();
loop {
let bytes_read = file
.read(&mut buf)
.unwrap_or_else(|e| panic!("error reading '{}': {}", path.display(), e));
if bytes_read == 0 {
break;
}
hasher.write(&buf[..bytes_read]);
}
let hash = hasher.finish();
format!(
"{}.{:x}.{}",
path.file_stem().unwrap().to_str().unwrap(),
hash,
path.extension().unwrap().to_str().unwrap()
)
}
fn make_url_path(url: &str) -> PathBuf {
let filename = url.replace(|c: char| !c.is_ascii_alphanumeric() && c != '.', "_");
Path::new(&std::env::var("OUT_DIR").unwrap()).join(filename)
}