You've already forked godot
							
							
				mirror of
				https://github.com/godotengine/godot.git
				synced 2025-11-03 11:50:27 +00:00 
			
		
		
		
	- Unify logic for loading editor/property/doc/extractable translations. - Replace legacy `TranslationServer` methods with translation domains for internal translations. - Only pre-create editor/property/doc translation domains in editor builds. - Prevent adding `null` translation. - Fixes potential loading of duplicated editor translations. - Add internal `has_translation_for_locale()` instead of calling `get_loaded_translations().has()`.
		
			
				
	
	
		
			155 lines
		
	
	
		
			4.7 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			155 lines
		
	
	
		
			4.7 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
"""Functions used to generate source files during build time"""
 | 
						|
 | 
						|
import os
 | 
						|
import os.path
 | 
						|
import subprocess
 | 
						|
import tempfile
 | 
						|
import uuid
 | 
						|
 | 
						|
import methods
 | 
						|
 | 
						|
 | 
						|
def doc_data_class_path_builder(target, source, env):
 | 
						|
    paths = dict(sorted(source[0].read().items()))
 | 
						|
    data = "\n".join([f'\t{{"{key}", "{value}"}},' for key, value in paths.items()])
 | 
						|
    with methods.generated_wrapper(str(target[0])) as file:
 | 
						|
        file.write(
 | 
						|
            f"""\
 | 
						|
struct _DocDataClassPath {{
 | 
						|
	const char *name;
 | 
						|
	const char *path;
 | 
						|
}};
 | 
						|
 | 
						|
inline constexpr int _doc_data_class_path_count = {len(paths)};
 | 
						|
inline constexpr _DocDataClassPath _doc_data_class_paths[{len(paths) + 1}] = {{
 | 
						|
	{data}
 | 
						|
	{{nullptr, nullptr}},
 | 
						|
}};
 | 
						|
"""
 | 
						|
        )
 | 
						|
 | 
						|
 | 
						|
def register_exporters_builder(target, source, env):
 | 
						|
    platforms = source[0].read()
 | 
						|
    exp_inc = "\n".join([f'#include "platform/{p}/export/export.h"' for p in platforms])
 | 
						|
    exp_reg = "\n\t".join([f"register_{p}_exporter();" for p in platforms])
 | 
						|
    exp_type = "\n\t".join([f"register_{p}_exporter_types();" for p in platforms])
 | 
						|
    with methods.generated_wrapper(str(target[0])) as file:
 | 
						|
        file.write(
 | 
						|
            f"""\
 | 
						|
#include "register_exporters.h"
 | 
						|
 | 
						|
{exp_inc}
 | 
						|
 | 
						|
void register_exporters() {{
 | 
						|
	{exp_reg}
 | 
						|
}}
 | 
						|
 | 
						|
void register_exporter_types() {{
 | 
						|
	{exp_type}
 | 
						|
}}
 | 
						|
"""
 | 
						|
        )
 | 
						|
 | 
						|
 | 
						|
def make_doc_header(target, source, env):
 | 
						|
    buffer = b"".join([methods.get_buffer(src) for src in map(str, source)])
 | 
						|
    decomp_size = len(buffer)
 | 
						|
    buffer = methods.compress_buffer(buffer)
 | 
						|
 | 
						|
    with methods.generated_wrapper(str(target[0])) as file:
 | 
						|
        file.write(f"""\
 | 
						|
inline constexpr const char *_doc_data_hash = "{hash(buffer)}";
 | 
						|
inline constexpr int _doc_data_compressed_size = {len(buffer)};
 | 
						|
inline constexpr int _doc_data_uncompressed_size = {decomp_size};
 | 
						|
inline constexpr const unsigned char _doc_data_compressed[] = {{
 | 
						|
	{methods.format_buffer(buffer, 1)}
 | 
						|
}};
 | 
						|
""")
 | 
						|
 | 
						|
 | 
						|
def make_translations(target, source, env):
 | 
						|
    target_h, target_cpp = str(target[0]), str(target[1])
 | 
						|
 | 
						|
    category = os.path.basename(target_h).split("_")[0]
 | 
						|
    sorted_paths = sorted([src.abspath for src in source], key=lambda path: os.path.splitext(os.path.basename(path))[0])
 | 
						|
 | 
						|
    xl_names = []
 | 
						|
    msgfmt = env.Detect("msgfmt")
 | 
						|
    if not msgfmt:
 | 
						|
        methods.print_warning("msgfmt not found, using .po files instead of .mo")
 | 
						|
 | 
						|
    with methods.generated_wrapper(target_cpp) as file:
 | 
						|
        for path in sorted_paths:
 | 
						|
            name = os.path.splitext(os.path.basename(path))[0]
 | 
						|
            # msgfmt erases non-translated messages, so avoid using it if exporting the POT.
 | 
						|
            if msgfmt and name != category:
 | 
						|
                mo_path = os.path.join(tempfile.gettempdir(), uuid.uuid4().hex + ".mo")
 | 
						|
                cmd = f"{msgfmt} {path} --no-hash -o {mo_path}"
 | 
						|
                try:
 | 
						|
                    subprocess.Popen(cmd, shell=True, stderr=subprocess.PIPE).communicate()
 | 
						|
                    buffer = methods.get_buffer(mo_path)
 | 
						|
                except OSError as e:
 | 
						|
                    methods.print_warning(
 | 
						|
                        "msgfmt execution failed, using .po file instead of .mo: path=%r; [%s] %s"
 | 
						|
                        % (path, e.__class__.__name__, e)
 | 
						|
                    )
 | 
						|
                    buffer = methods.get_buffer(path)
 | 
						|
                finally:
 | 
						|
                    try:
 | 
						|
                        if os.path.exists(mo_path):
 | 
						|
                            os.remove(mo_path)
 | 
						|
                    except OSError as e:
 | 
						|
                        # Do not fail the entire build if it cannot delete a temporary file.
 | 
						|
                        methods.print_warning(
 | 
						|
                            "Could not delete temporary .mo file: path=%r; [%s] %s" % (mo_path, e.__class__.__name__, e)
 | 
						|
                        )
 | 
						|
            else:
 | 
						|
                buffer = methods.get_buffer(path)
 | 
						|
                if name == category:
 | 
						|
                    name = "source"
 | 
						|
 | 
						|
            decomp_size = len(buffer)
 | 
						|
            buffer = methods.compress_buffer(buffer)
 | 
						|
 | 
						|
            file.write(f"""\
 | 
						|
inline constexpr const unsigned char _{category}_translation_{name}_compressed[] = {{
 | 
						|
	{methods.format_buffer(buffer, 1)}
 | 
						|
}};
 | 
						|
 | 
						|
""")
 | 
						|
 | 
						|
            xl_names.append([name, len(buffer), decomp_size])
 | 
						|
 | 
						|
        file.write(f"""\
 | 
						|
#include "{target_h}"
 | 
						|
 | 
						|
const EditorTranslationList _{category}_translations[] = {{
 | 
						|
""")
 | 
						|
 | 
						|
        for x in xl_names:
 | 
						|
            file.write(f'\t{{ "{x[0]}", {x[1]}, {x[2]}, _{category}_translation_{x[0]}_compressed }},\n')
 | 
						|
 | 
						|
        file.write("""\
 | 
						|
	{ nullptr, 0, 0, nullptr },
 | 
						|
};
 | 
						|
""")
 | 
						|
 | 
						|
    with methods.generated_wrapper(target_h) as file:
 | 
						|
        file.write(f"""\
 | 
						|
 | 
						|
#ifndef EDITOR_TRANSLATION_LIST
 | 
						|
#define EDITOR_TRANSLATION_LIST
 | 
						|
 | 
						|
struct EditorTranslationList {{
 | 
						|
	const char* lang;
 | 
						|
	int comp_size;
 | 
						|
	int uncomp_size;
 | 
						|
	const unsigned char* data;
 | 
						|
}};
 | 
						|
 | 
						|
#endif // EDITOR_TRANSLATION_LIST
 | 
						|
 | 
						|
extern const EditorTranslationList _{category}_translations[];
 | 
						|
""")
 |