aboutsummaryrefslogtreecommitdiff
path: root/src
diff options
context:
space:
mode:
Diffstat (limited to 'src')
-rw-r--r--src/main.rs106
1 files changed, 80 insertions, 26 deletions
diff --git a/src/main.rs b/src/main.rs
index 533f157..26e3980 100644
--- a/src/main.rs
+++ b/src/main.rs
@@ -46,10 +46,8 @@ fn main() {
}
Cmd::ParseJlptVocab => {
let kanji_levels = read_kanji_levels().expect("read_kanji_levels");
- let all_kanji = Charset::from_iter(kanji_levels
- .iter()
- .map(|(_, c)| c.chars())
- .flatten());
+ let all_kanji =
+ Charset::from_iter(kanji_levels.iter().map(|(_, c)| c.chars()).flatten());
parse_jlpt_vocab(&all_kanji).expect("error");
}
Cmd::New { truncate, count } => {
@@ -324,29 +322,59 @@ struct JlptVocab {
impl JlptVocab {
fn to_string(&self) -> String {
- format!("{}\t{}\t{}\t{}\t{}",
- self.level,
- self.chars.to_string(),
- self.kanji,
- self.kana,
- self.en)
+ format!(
+ "{}\t{}\t{}\t{}\t{}",
+ self.level,
+ self.chars.to_string(),
+ self.kanji,
+ self.kana,
+ self.en
+ )
}
}
fn parse_jlpt_vocab(all_kanji: &Charset) -> Result<()> {
let mut vocab = vec![];
- vocab.extend(parse_jlpt_vocab_combined("data/n5_vocab.txt", "N4", all_kanji)?);
- vocab.extend(parse_jlpt_vocab_split("data/n4_vocab_hiragana.txt", "data/n4_vocab_eng.txt", "N3", all_kanji)?);
- vocab.extend(parse_jlpt_vocab_split("data/n3_vocab_hiragana.txt", "data/n3_vocab_eng.txt", "N2a", all_kanji)?);
- vocab.extend(parse_jlpt_vocab_split("data/n2_vocab_hiragana.txt", "data/n2_vocab_eng.txt", "N2b", all_kanji)?);
- vocab.extend(parse_jlpt_vocab_split("data/n1_vocab_hiragana.txt", "data/n1_vocab_eng.txt", "N1", all_kanji)?);
+ vocab.extend(parse_jlpt_vocab_combined(
+ "data/n5_vocab.txt",
+ "N4",
+ all_kanji,
+ )?);
+ vocab.extend(parse_jlpt_vocab_split(
+ "data/n4_vocab_hiragana.txt",
+ "data/n4_vocab_eng.txt",
+ "N3",
+ all_kanji,
+ )?);
+ vocab.extend(parse_jlpt_vocab_split(
+ "data/n3_vocab_hiragana.txt",
+ "data/n3_vocab_eng.txt",
+ "N2a",
+ all_kanji,
+ )?);
+ vocab.extend(parse_jlpt_vocab_split(
+ "data/n2_vocab_hiragana.txt",
+ "data/n2_vocab_eng.txt",
+ "N2b",
+ all_kanji,
+ )?);
+ vocab.extend(parse_jlpt_vocab_split(
+ "data/n1_vocab_hiragana.txt",
+ "data/n1_vocab_eng.txt",
+ "N1",
+ all_kanji,
+ )?);
for v in vocab.iter() {
println!("{}", v.to_string());
}
Ok(())
}
-fn parse_jlpt_vocab_combined(file: &str, level: &str, all_kanji: &Charset) -> Result<Vec<JlptVocab>> {
+fn parse_jlpt_vocab_combined(
+ file: &str,
+ level: &str,
+ all_kanji: &Charset,
+) -> Result<Vec<JlptVocab>> {
let lines = jlpt_vocab_read_file(file)?;
let mut ret = vec![];
for (kanji, answer) in lines {
@@ -360,14 +388,19 @@ fn parse_jlpt_vocab_combined(file: &str, level: &str, all_kanji: &Charset) -> Re
chars: Charset::new(kanji).inter(all_kanji),
kanji: kanji.to_string(),
kana: kana.to_string(),
- en: eng.to_string()
+ en: eng.to_string(),
});
}
}
Ok(ret)
}
-fn parse_jlpt_vocab_split(kana_file: &str, eng_file: &str, level: &str, all_kanji: &Charset) -> Result<Vec<JlptVocab>> {
+fn parse_jlpt_vocab_split(
+ kana_file: &str,
+ eng_file: &str,
+ level: &str,
+ all_kanji: &Charset,
+) -> Result<Vec<JlptVocab>> {
let eng_lines = jlpt_vocab_read_file(eng_file)?
.into_iter()
.collect::<HashMap<String, String>>();
@@ -383,7 +416,7 @@ fn parse_jlpt_vocab_split(kana_file: &str, eng_file: &str, level: &str, all_kanj
chars: Charset::new(kanji).inter(all_kanji),
kanji: kanji.to_string(),
kana: kana.to_string(),
- en: eng.to_string()
+ en: eng.to_string(),
});
}
}
@@ -397,8 +430,7 @@ fn jlpt_vocab_read_file(file: &str) -> Result<Vec<(String, String)>> {
let file = fs::File::open(file)?;
let mut ret = vec![];
for line in io::BufReader::new(file).lines() {
- let line = line?.replace("<br>", "\n")
- .replace("</span>", "");
+ let line = line?.replace("<br>", "\n").replace("</span>", "");
let line = re.replace_all(&line, "");
if let Some((a, b)) = line.split_once('|') {
ret.push((a.trim().to_string(), b.trim().to_string()));
@@ -1034,7 +1066,15 @@ fn add_vocab(all_batches: &mut [Batch], vocab: &[JlptVocab]) {
.filter(|v| v.chars.inter_len(&batch.chars) > 0)
.filter(|v| match_level(batch, &v.level))
.filter(|v| v.chars.diff(&done_after).len() == 0)
- .filter(|v| !all_batches[i..std::cmp::min(all_batches.len(), i+10)].iter().any(|b| b.examples.iter().any(|ex| ex.ja.contains(&v.kanji) || ex.expl.contains(&v.kanji))))
+ .filter(|v| {
+ !all_batches[i..std::cmp::min(all_batches.len(), i + 10)]
+ .iter()
+ .any(|b| {
+ b.examples
+ .iter()
+ .any(|ex| ex.ja.contains(&v.kanji) || ex.expl.contains(&v.kanji))
+ })
+ })
.cloned()
.collect::<Vec<_>>();
extra_vocab.push(batch_extra_vocab);
@@ -1156,18 +1196,32 @@ fn format_batch_aux<'a>(
}
writeln!(f, "<hr />")?;
- writeln!(f, r#"<details><summary>Extra vocabulary (this level)</summary>"#)?;
+ writeln!(
+ f,
+ r#"<details><summary>Extra vocabulary (this level)</summary>"#
+ )?;
for v in batch.extra_vocab.iter() {
if batch.level.contains(&v.level) {
- writeln!(f, r#"<p>({}) {} [{}] {}</p>"#, v.level, v.kanji, v.kana, v.en)?;
+ writeln!(
+ f,
+ r#"<p>({}) {} [{}] {}</p>"#,
+ v.level, v.kanji, v.kana, v.en
+ )?;
}
}
writeln!(f, r#"</details>"#)?;
if !batch.level.contains("N4") {
- writeln!(f, r#"<details><summary>Extra vocabulary (previous levels)</summary>"#)?;
+ writeln!(
+ f,
+ r#"<details><summary>Extra vocabulary (previous levels)</summary>"#
+ )?;
for v in batch.extra_vocab.iter() {
if !batch.level.contains(&v.level) {
- writeln!(f, r#"<p>({}) {} [{}] {}</p>"#, v.level, v.kanji, v.kana, v.en)?;
+ writeln!(
+ f,
+ r#"<p>({}) {} [{}] {}</p>"#,
+ v.level, v.kanji, v.kana, v.en
+ )?;
}
}
writeln!(f, r#"</details>"#)?;