mirror of
https://github.com/RoboSats/robosats.git
synced 2024-12-14 11:26:24 +00:00
6a1c8c32ef
Changed handcrafted.py to use a OrderedDict instead of a dict (guarantee the order independant of the python version) Moved the value of "../../src" to the variable "src_path" to improve readability, and reuse the value again later in code Changed the "Look for all matching i18n keys" part of code to order the dirs and file in a independant OS way - alphabetical order using the "sort" func of list object Changed the "Look for all matching i18n keys" part of code to collect all phrases in a OrderedDict (guarantee the order independant of the python version) Changed the list indexing in filepath var (filepath[10:]) to the use of os.path.relpath(filepath, src_path) func to improve readability Changed all_phrases to a OrderedDict (guarantee the order independant of the python version) and use update method of OrderedDict to add the files's phrases and handcrafted phrases to the OrderedDict in order Changed the loop of matches to inside the if statement (if len(matches) > 0) to improve readability - if the "if statement" is false, the loop was never executed Changed the "Load existing locale dics and replace keys" part of code to use OrderedDict in the new_phrases var to guarantee the order of the keys insert in json.dump in all OS Changed the "Load existing locale dics and replace keys" part of code to not substitute the keys "#{counter}" to guarantee the order independant of the json file of the translation.
63 lines
2.1 KiB
Python
63 lines
2.1 KiB
Python
"""
|
|
This script looks for any t('phrase') in every frontend source files .src/
|
|
extracts the strings into a new dictionary (en.json), and updates every other
|
|
locale dictionary by matching existing translations.
|
|
"""
|
|
|
|
import json
|
|
import os
|
|
import re
|
|
from collections import OrderedDict
|
|
|
|
from handcrafted import phrases
|
|
|
|
prefix = r"[^a-zA-Z]t\s*\(\s*\'"
|
|
suffix = r"(?:\'\,\s*\)|\'\,*\)|\'\)|\'\,\s*\{)"
|
|
match = r"(.*?[a-zA-Z].*?)"
|
|
pattern = f"{prefix}{match}{suffix}"
|
|
|
|
extensions = [".ts", ".tsx", ".js", "jsx"]
|
|
strings_dict = OrderedDict()
|
|
src_path = "../../src"
|
|
counter = 1
|
|
|
|
# Look for all matching i18n keys in src_path
|
|
for root, dirs, files in os.walk(src_path):
|
|
dirs.sort()
|
|
files.sort()
|
|
for file in files:
|
|
if file.endswith(tuple(extensions)):
|
|
filepath = os.path.join(root, file)
|
|
with open(filepath, "r", encoding="utf-8") as f:
|
|
contents = f.read()
|
|
matches = sorted(re.findall(pattern, contents))
|
|
if len(matches) > 0:
|
|
rel_filepath_in_src = os.path.relpath(filepath, src_path)
|
|
strings_dict[f"#{counter}"] = f"Phrases in {rel_filepath_in_src}"
|
|
counter += 1
|
|
for match in matches:
|
|
strings_dict[match] = match
|
|
|
|
all_phrases = OrderedDict()
|
|
all_phrases.update(strings_dict)
|
|
all_phrases.update(phrases)
|
|
|
|
# Load existing locale dics and replace keys
|
|
locales = [f for f in os.listdir(".") if f.endswith(".json")]
|
|
for locale in locales:
|
|
new_phrases = OrderedDict()
|
|
with open(locale, "r", encoding="utf-8") as f:
|
|
old_phrases = json.load(f)
|
|
for key in all_phrases.keys():
|
|
# update dictionary with new keys on /src/, but ignore the counter of files keys
|
|
if key in old_phrases and not re.match(r"^#\d+$", key):
|
|
new_phrases[key] = old_phrases[key]
|
|
else:
|
|
new_phrases[key] = all_phrases[key]
|
|
|
|
with open(locale, "w", encoding="utf-8") as f:
|
|
json.dump(new_phrases, f, ensure_ascii=False)
|
|
|
|
with open("./collected_phrases.json", "w", encoding="utf-8") as f:
|
|
json.dump(all_phrases, f, ensure_ascii=False)
|