Add code to use OpenAI API to do transalations
This commit is contained in:
861
sm2gen.py
861
sm2gen.py
@@ -9,6 +9,9 @@ import os
|
||||
import pkg_resources
|
||||
from datetime import datetime
|
||||
import xml.etree.ElementTree as ET
|
||||
from openai import OpenAI
|
||||
import configparser
|
||||
import json
|
||||
|
||||
#
|
||||
#To Do
|
||||
@@ -16,113 +19,116 @@ import xml.etree.ElementTree as ET
|
||||
# 2. Deal with ? and ! as sentence ends in text-format
|
||||
# 3. Make lexical scan work with double or single quotes (if possible)
|
||||
|
||||
SME2Gen_version = '0.7'
|
||||
SME2Gen_version = '0.8'
|
||||
json5_dict: dict = {}
|
||||
json5_html_list: list = []
|
||||
# Define the path to the INI file
|
||||
ini_file_path = os.path.expanduser('~/.smegit/conf')
|
||||
OPENAI_API_KEY = ""
|
||||
|
||||
def parse_json(json_obj, prefix=''):
|
||||
structured_list = []
|
||||
if isinstance(json_obj, dict):
|
||||
for k, v in json_obj.items():
|
||||
new_key = f"{prefix}.{k}" if prefix else k
|
||||
structured_list.extend(parse_json(v, new_key))
|
||||
elif isinstance(json_obj, list):
|
||||
for i, v in enumerate(json_obj):
|
||||
new_key = f"{prefix}[{i}]"
|
||||
structured_list.extend(parse_json(v, new_key))
|
||||
else:
|
||||
structured_list.append(f"{prefix}: {json_obj}")
|
||||
return structured_list
|
||||
structured_list = []
|
||||
if isinstance(json_obj, dict):
|
||||
for k, v in json_obj.items():
|
||||
new_key = f"{prefix}.{k}" if prefix else k
|
||||
structured_list.extend(parse_json(v, new_key))
|
||||
elif isinstance(json_obj, list):
|
||||
for i, v in enumerate(json_obj):
|
||||
new_key = f"{prefix}[{i}]"
|
||||
structured_list.extend(parse_json(v, new_key))
|
||||
else:
|
||||
structured_list.append(f"{prefix}: {json_obj}")
|
||||
return structured_list
|
||||
|
||||
def json5_to_list(filename):
|
||||
with open(filename, 'r') as file:
|
||||
data = json5.load(file)
|
||||
return parse_json(data)
|
||||
with open(filename, 'r') as file:
|
||||
data = json5.load(file)
|
||||
return parse_json(data)
|
||||
|
||||
def json5_to_pandas(filename):
|
||||
with open(filename, 'r') as file:
|
||||
data = json5.load(file)
|
||||
print (data)
|
||||
return data.json_normalize(data)
|
||||
with open(filename, 'r') as file:
|
||||
data = json5.load(file)
|
||||
print (data)
|
||||
return data.json_normalize(data)
|
||||
|
||||
def json5_to_dict(filename):
|
||||
with open(filename, 'r') as file:
|
||||
data = json5.load(file)
|
||||
return data
|
||||
with open(filename, 'r') as file:
|
||||
data = json5.load(file)
|
||||
return data
|
||||
|
||||
def rec_print(data, prefix=''):
|
||||
# Check if this item is a dictionary.
|
||||
if isinstance(data, dict):
|
||||
for key, val in data.items():
|
||||
rec_print(val, f"{prefix}.{key}")
|
||||
# Check if this item is a list.
|
||||
elif isinstance(data, list):
|
||||
for idx, val in enumerate(data):
|
||||
rec_print(val, f"{prefix}[{idx}]")
|
||||
# If neither, it's a basic type.
|
||||
else:
|
||||
print(f"{prefix}: {data}")
|
||||
|
||||
# Check if this item is a dictionary.
|
||||
if isinstance(data, dict):
|
||||
for key, val in data.items():
|
||||
rec_print(val, f"{prefix}.{key}")
|
||||
# Check if this item is a list.
|
||||
elif isinstance(data, list):
|
||||
for idx, val in enumerate(data):
|
||||
rec_print(val, f"{prefix}[{idx}]")
|
||||
# If neither, it's a basic type.
|
||||
else:
|
||||
print(f"{prefix}: {data}")
|
||||
|
||||
def find_item(nested_dict, target_key):
|
||||
for key, val in nested_dict.items():
|
||||
if key == target_key:
|
||||
return val
|
||||
elif isinstance(val, dict):
|
||||
result = find_item(val, target_key)
|
||||
if result is not None:
|
||||
return result
|
||||
for key, val in nested_dict.items():
|
||||
if key == target_key:
|
||||
return val
|
||||
elif isinstance(val, dict):
|
||||
result = find_item(val, target_key)
|
||||
if result is not None:
|
||||
return result
|
||||
|
||||
def find_dicts_with_key(data, target_key):
|
||||
results = []
|
||||
if isinstance(data, dict):
|
||||
if target_key in data:
|
||||
results.append(data)
|
||||
for val in data.values():
|
||||
if isinstance(val, (dict, list)):
|
||||
results.extend(find_dicts_with_key(val, target_key))
|
||||
elif isinstance(data, list):
|
||||
for item in data:
|
||||
if isinstance(item, (dict, list)):
|
||||
results.extend(find_dicts_with_key(item, target_key))
|
||||
return results
|
||||
|
||||
results = []
|
||||
if isinstance(data, dict):
|
||||
if target_key in data:
|
||||
results.append(data)
|
||||
for val in data.values():
|
||||
if isinstance(val, (dict, list)):
|
||||
results.extend(find_dicts_with_key(val, target_key))
|
||||
elif isinstance(data, list):
|
||||
for item in data:
|
||||
if isinstance(item, (dict, list)):
|
||||
results.extend(find_dicts_with_key(item, target_key))
|
||||
return results
|
||||
|
||||
def find_values_with_key(data, target_key):
|
||||
results = []
|
||||
if isinstance(data, dict):
|
||||
if target_key in data:
|
||||
results.append(data[target_key])
|
||||
for val in data.values():
|
||||
if isinstance(val, (dict, list)):
|
||||
results.extend(find_values_with_key(val, target_key))
|
||||
elif isinstance(data, list):
|
||||
for item in data:
|
||||
if isinstance(item, (dict, list)):
|
||||
results.extend(find_values_with_key(item, target_key))
|
||||
return results
|
||||
|
||||
results = []
|
||||
if isinstance(data, dict):
|
||||
if target_key in data:
|
||||
results.append(data[target_key])
|
||||
for val in data.values():
|
||||
if isinstance(val, (dict, list)):
|
||||
results.extend(find_values_with_key(val, target_key))
|
||||
elif isinstance(data, list):
|
||||
for item in data:
|
||||
if isinstance(item, (dict, list)):
|
||||
results.extend(find_values_with_key(item, target_key))
|
||||
return results
|
||||
|
||||
def lint_json5(filename):
|
||||
try:
|
||||
with open(filename, 'r') as file:
|
||||
data = file.read()
|
||||
json5.loads(data)
|
||||
print(f"{filename} as JSON5 data is valid")
|
||||
except Exception as e:
|
||||
print(f"{filename} as JSON5 data is invalid")
|
||||
print("Error:", str(e))
|
||||
sys.exit()
|
||||
try:
|
||||
with open(filename, 'r') as file:
|
||||
data = file.read()
|
||||
json5.loads(data)
|
||||
print(f"{filename} as JSON5 data is valid")
|
||||
except Exception as e:
|
||||
print(f"{filename} as JSON5 data is invalid")
|
||||
print("Error:", str(e))
|
||||
sys.exit()
|
||||
|
||||
|
||||
def flatten_hash_of_lists(hash_of_lists):
|
||||
flattened = {}
|
||||
for key, value in hash_of_lists.items():
|
||||
if isinstance(value, list):
|
||||
for i, item in enumerate(value):
|
||||
new_key = f"{key}_{i}" # Appending index to the key to maintain uniqueness
|
||||
flattened[new_key] = item
|
||||
else:
|
||||
flattened[key] = value
|
||||
return flattened
|
||||
|
||||
flattened = {}
|
||||
for key, value in hash_of_lists.items():
|
||||
if isinstance(value, list):
|
||||
for i, item in enumerate(value):
|
||||
new_key = f"{key}_{i}" # Appending index to the key to maintain uniqueness
|
||||
flattened[new_key] = item
|
||||
else:
|
||||
flattened[key] = value
|
||||
return flattened
|
||||
|
||||
def hl(keyname):
|
||||
# Return highest level value for the keyname
|
||||
if keyname in json5_dict:
|
||||
@@ -132,53 +138,53 @@ def hl(keyname):
|
||||
return 'None'
|
||||
|
||||
def get_all_routes():
|
||||
route_list = [html_block.get('route') for html_block in json5_dict.get('html', [])]
|
||||
return route_list
|
||||
route_list = [html_block.get('route') for html_block in json5_dict.get('html', [])]
|
||||
return route_list
|
||||
|
||||
def lc_get_all_routes():
|
||||
# All routes in lower case
|
||||
route_list = [html_block.get('route').lower() for html_block in json5_dict.get('html', [])]
|
||||
return route_list
|
||||
|
||||
# All routes in lower case
|
||||
route_list = [html_block.get('route').lower() for html_block in json5_dict.get('html', [])]
|
||||
return route_list
|
||||
|
||||
def has_file_been_modified(file_path):
|
||||
# Get the file's creation time and last modification time in Unix timestamp
|
||||
creation_time = os.path.getctime(file_path)
|
||||
last_modification_time = os.path.getmtime(file_path)
|
||||
print(f"{creation_time}*{last_modification_time}")
|
||||
quit()
|
||||
# Get the file's creation time and last modification time in Unix timestamp
|
||||
creation_time = os.path.getctime(file_path)
|
||||
last_modification_time = os.path.getmtime(file_path)
|
||||
print(f"{creation_time}*{last_modification_time}")
|
||||
quit()
|
||||
|
||||
# Compare the creation time and last modification time
|
||||
if creation_time < last_modification_time:
|
||||
return True # File has been modified after creation
|
||||
else:
|
||||
return False # File has not been modified after creation
|
||||
# Compare the creation time and last modification time
|
||||
if creation_time < last_modification_time:
|
||||
return True # File has been modified after creation
|
||||
else:
|
||||
return False # File has not been modified after creation
|
||||
|
||||
def parse_xml_to_dict(xml_file):
|
||||
# Parse the XML file
|
||||
tree = ET.parse(xml_file)
|
||||
root = tree.getroot()
|
||||
# Parse the XML file
|
||||
tree = ET.parse(xml_file)
|
||||
root = tree.getroot()
|
||||
|
||||
xml_dict = {} # Initialize an empty dictionary to store the data
|
||||
xml_dict = {} # Initialize an empty dictionary to store the data
|
||||
|
||||
# Iterate through the XML tree and extract data
|
||||
for elem in root:
|
||||
tag = elem.tag
|
||||
if elem.text:
|
||||
xml_dict[tag] = elem.text
|
||||
else:
|
||||
cdata_content = elem.find('.//').text # Extract CDATA text
|
||||
xml_dict[tag] = cdata_content
|
||||
# Iterate through the XML tree and extract data
|
||||
for elem in root:
|
||||
tag = elem.tag
|
||||
if elem.text:
|
||||
xml_dict[tag] = elem.text
|
||||
else:
|
||||
cdata_content = elem.find('.//').text # Extract CDATA text
|
||||
xml_dict[tag] = cdata_content
|
||||
|
||||
return xml_dict
|
||||
return xml_dict
|
||||
|
||||
def deduplicate_array(arr):
|
||||
# Convert the array to a set to remove duplicates
|
||||
unique_set = set(arr)
|
||||
|
||||
# Convert the set back to a list to maintain the order
|
||||
deduplicated_list = list(unique_set)
|
||||
|
||||
return deduplicated_list
|
||||
# Convert the array to a set to remove duplicates
|
||||
unique_set = set(arr)
|
||||
|
||||
# Convert the set back to a list to maintain the order
|
||||
deduplicated_list = list(unique_set)
|
||||
|
||||
return deduplicated_list
|
||||
|
||||
def get_db_fields():
|
||||
return []
|
||||
@@ -187,308 +193,351 @@ def get_table_control_data():
|
||||
return find_values_with_key(json5_html_list,'TableControl')
|
||||
|
||||
def format_text(text):
|
||||
#
|
||||
# Turn a piece of text into something a bit better formatted - spaces after full stop and comma and also capitalise sentences
|
||||
#
|
||||
# Create a list to hold the formatted sentences
|
||||
formatted_sentences = []
|
||||
# Split the text into sentences
|
||||
sentences = text.split(".")
|
||||
words = sentences[0].split(" ")
|
||||
#print(len(sentences))
|
||||
# Deal with one capitalised word
|
||||
if sentences[0].isupper() and len(sentences) == 1 and len(words) == 1:
|
||||
return sentences[0].capitalize()
|
||||
else:
|
||||
for sentence in sentences:
|
||||
#print(sentence)
|
||||
# and splt into sub phrases, based on comma
|
||||
formatted_phrases = []
|
||||
phrases = sentence.split(",")
|
||||
#
|
||||
# Turn a piece of text into something a bit better formatted - spaces after full stop and comma and also capitalise sentences
|
||||
#
|
||||
# Create a list to hold the formatted sentences
|
||||
formatted_sentences = []
|
||||
# Split the text into sentences
|
||||
sentences = text.split(".")
|
||||
words = sentences[0].split(" ")
|
||||
#print(len(sentences))
|
||||
# Deal with one capitalised word
|
||||
if sentences[0].isupper() and len(sentences) == 1 and len(words) == 1:
|
||||
return sentences[0].capitalize()
|
||||
else:
|
||||
for sentence in sentences:
|
||||
#print(sentence)
|
||||
# and splt into sub phrases, based on comma
|
||||
formatted_phrases = []
|
||||
phrases = sentence.split(",")
|
||||
|
||||
for phrase in phrases:
|
||||
#print(phrase)
|
||||
phrase = phrase.lstrip()
|
||||
formatted_words = []
|
||||
words = phrase.split(' ')
|
||||
|
||||
for i,word in enumerate(words):
|
||||
#print(i,word)
|
||||
# Check if the word is fully uppercase or not the first
|
||||
word = word.lstrip()
|
||||
if word.isupper() or i != 0:
|
||||
formatted_words.append(word)
|
||||
else:
|
||||
# Capitalize the word
|
||||
formatted_words.append(word.capitalize())
|
||||
for phrase in phrases:
|
||||
#print(phrase)
|
||||
phrase = phrase.lstrip()
|
||||
formatted_words = []
|
||||
words = phrase.split(' ')
|
||||
|
||||
for i,word in enumerate(words):
|
||||
#print(i,word)
|
||||
# Check if the word is fully uppercase or not the first
|
||||
word = word.lstrip()
|
||||
if word.isupper() or i != 0:
|
||||
formatted_words.append(word)
|
||||
else:
|
||||
# Capitalize the word
|
||||
formatted_words.append(word.capitalize())
|
||||
|
||||
# Join the formatted words back together for this phrase
|
||||
formatted_phrase = ' '.join(formatted_words).lstrip()
|
||||
formatted_phrases.append(formatted_phrase)
|
||||
# and sentence
|
||||
formatted_sentence = ", ".join(formatted_phrases).lstrip()
|
||||
formatted_sentences.append(formatted_sentence)
|
||||
# Join the formatted sentences back together
|
||||
formatted_text = ". ".join(formatted_sentences).lstrip()
|
||||
return formatted_text
|
||||
|
||||
def get_completion(prompt):
|
||||
response = client.chat.completions.create(model="gpt-4o",
|
||||
messages=[{'role':"user", 'content':prompt}],
|
||||
max_tokens=100)
|
||||
# Extracting the content message from the JSON data
|
||||
response_data = response.to_dict()
|
||||
content_message = response_data["choices"][0]["message"]["content"]
|
||||
return content_message.strip()
|
||||
|
||||
def get_translation(message="Hello",language="french"):
|
||||
prompt = f"Translate the following text from English to {language}: {message} :suitable for use in a Server manager application"
|
||||
translated_message = get_completion(prompt)
|
||||
return translated_message
|
||||
|
||||
def convert_lex_to_dict(pairs_string):
|
||||
# Splitting pairs string by comma and newline, excluding empty lines
|
||||
pairs_list = [line for line in pairs_string.split(",\n") if line.strip()]
|
||||
data_dict = {pair.split(" => ")[0].strip("'"): pair.split(" => ")[1].strip("'") for pair in pairs_list}
|
||||
formatted_dict = [{"id": key, "text": value} for key, value in data_dict.items()]
|
||||
return formatted_dict
|
||||
|
||||
# Join the formatted words back together for this phrase
|
||||
formatted_phrase = ' '.join(formatted_words).lstrip()
|
||||
formatted_phrases.append(formatted_phrase)
|
||||
# and sentence
|
||||
formatted_sentence = ", ".join(formatted_phrases).lstrip()
|
||||
formatted_sentences.append(formatted_sentence)
|
||||
# Join the formatted sentences back together
|
||||
formatted_text = ". ".join(formatted_sentences).lstrip()
|
||||
return formatted_text
|
||||
|
||||
if __name__ == "__main__":
|
||||
try:
|
||||
chameleon_version = pkg_resources.get_distribution("Chameleon").version
|
||||
except pkg_resources.DistributionNotFound:
|
||||
chameleon_version = "Version information not available"
|
||||
python_version = sys.version
|
||||
python_version = python_version[:8]
|
||||
current_datetime = datetime.now()
|
||||
formatted_datetime = current_datetime.strftime("%Y-%m-%d %H:%M")
|
||||
try:
|
||||
chameleon_version = pkg_resources.get_distribution("Chameleon").version
|
||||
except pkg_resources.DistributionNotFound:
|
||||
chameleon_version = "Version information not available"
|
||||
python_version = sys.version
|
||||
python_version = python_version[:8]
|
||||
current_datetime = datetime.now()
|
||||
formatted_datetime = current_datetime.strftime("%Y-%m-%d %H:%M")
|
||||
|
||||
strVersion = "SME2Gen version:"+SME2Gen_version+" Chameleon version:"+chameleon_version+" On Python:"+python_version+" at "+formatted_datetime
|
||||
strVersion = "SM2Gen version:"+SME2Gen_version+" Chameleon version:"+chameleon_version+" On Python:"+python_version+" at "+formatted_datetime
|
||||
|
||||
json5_dict: dict = {}
|
||||
json5_html_list: list = []
|
||||
json5_dict: dict = {}
|
||||
json5_html_list: list = []
|
||||
|
||||
print(f"SM2 code from JSON5 - {strVersion}")
|
||||
|
||||
print(f"SM2 code from JSON5 - {strVersion}")
|
||||
home_dir = "/home/brianr/clients/SM2/SM2Gen/"
|
||||
json_filename = f"{home_dir}nfsshare.json5"
|
||||
|
||||
# read open ai key from ini file
|
||||
# Check if the file exists
|
||||
if os.path.exists(ini_file_path):
|
||||
# Create a configparser object and read the file
|
||||
config = configparser.ConfigParser()
|
||||
config.read(ini_file_path)
|
||||
|
||||
filename = '/home/brianr/clients/SM2/SM2Gen/nfsshare.json5'
|
||||
|
||||
# Command line parameters
|
||||
parser = argparse.ArgumentParser(description="SM2Gen")
|
||||
parser.add_argument('-f', '--filename', help='Specify a filename for the JSON5 file', default=filename)
|
||||
parser.add_argument('-nco', '--noController', help='Stop it creating a controller file', default="no")
|
||||
parser.add_argument('-nh', '--noHtml', help='Stop it creating html files(s)', default="no")
|
||||
parser.add_argument('-nl', '--noLang', help='Stop it creating language localise files(s)', default="no")
|
||||
parser.add_argument('-ncu', '--noCust', help='Stop it creating Custom controller file', default="no")
|
||||
args = parser.parse_args()
|
||||
filename = args.filename
|
||||
print(f"JSON5 from {filename} with noController={args.noController}, noHtml={args.noHtml} and noLang={args.noLang}") #Not yet activated
|
||||
# Read the value of "OPENAI_API_KEY"
|
||||
if 'OPENAI_API_KEY' in config['smegit']:
|
||||
OPENAI_API_KEY = config['smegit']['OPENAI_API_KEY']
|
||||
#print("API Key:", OPENAI_API_KEY)
|
||||
client = OpenAI(api_key=OPENAI_API_KEY)
|
||||
else:
|
||||
print("OPENAI_API_KEY not found in the configuration file.")
|
||||
else:
|
||||
print("Configuration file not found at:", file_path)
|
||||
|
||||
# Command line parameters - not in use
|
||||
parser = argparse.ArgumentParser(description="SM2Gen")
|
||||
parser.add_argument('-f', '--filename', help='Specify a filename for the JSON5 file', default=json_filename)
|
||||
parser.add_argument('-nco', '--noController', help='Stop it creating a controller file', default="no")
|
||||
parser.add_argument('-nh', '--noHtml', help='Stop it creating html files(s)', default="no")
|
||||
parser.add_argument('-nl', '--noLang', help='Stop it creating language localise files(s)', default="no")
|
||||
parser.add_argument('-ncu', '--noCust', help='Stop it creating Custom controller file', default="no")
|
||||
args = parser.parse_args()
|
||||
json_filename = args.filename
|
||||
print(f"JSON5 from {json_filename} with noController={args.noController}, noHtml={args.noHtml} and noLang={args.noLang}") #Not yet activated
|
||||
|
||||
# check syntax of JSON5
|
||||
lint_json5(filename);
|
||||
|
||||
# Get dict of it all
|
||||
json5_dict = json5_to_dict(filename)
|
||||
|
||||
# Get dict of just the html bit
|
||||
json5_html_list = json5_dict['html']
|
||||
|
||||
#Identify message
|
||||
print(f"\nGenerating mojo panels for {hl('PackageName')}")
|
||||
print( "-----------------------------------")
|
||||
|
||||
# Routes for each panel
|
||||
routes = get_all_routes();
|
||||
lc_routes =lc_get_all_routes();
|
||||
|
||||
#File names
|
||||
controller_file = 'Targets/'+hl('PackageName')+'.pm'
|
||||
custom_controller_file = 'Targets/'+hl('PackageName')+'-Custom.pm'
|
||||
|
||||
# see if it has been modified by developer - this did not work - the modified by was the same as the creation time/date
|
||||
#if has_file_been_modified(custom_controller_file):
|
||||
# custom_controller_file = custom_controller_file+'.new'
|
||||
|
||||
# Call it .new if one is already there (and may have been editted by the developer)
|
||||
if os.path.exists(custom_controller_file):
|
||||
custom_controller_file = custom_controller_file+'.new'
|
||||
layout_file = 'Targets/'+hl('PackageName')+'.html.ep'
|
||||
partial_files = list()
|
||||
for panel in routes:
|
||||
partial_files.append('Targets/_'+hl('prefix')+"_"+panel+'.html.ep')
|
||||
print(partial_files)
|
||||
lex_file = 'Targets/'+hl('PackageName').lower()+'_en.lex'
|
||||
tablecontrols = get_table_control_data() #arrays of hashes used to drive rows in tables
|
||||
|
||||
#Generate controller file
|
||||
try:
|
||||
controller_template = PageTemplateFile("Templates/controller.pm.tem")
|
||||
dbentries = get_db_fields() #Params which correspond to Db fields
|
||||
try:
|
||||
controller_perl = controller_template.render(version=strVersion,
|
||||
lint_json5(json_filename);
|
||||
|
||||
# Get dict of it all
|
||||
json5_dict = json5_to_dict(json_filename)
|
||||
|
||||
# Get dict of just the html bit
|
||||
json5_html_list = json5_dict['html']
|
||||
|
||||
#Identify message
|
||||
print(f"\nGenerating mojo panels for {hl('PackageName')}")
|
||||
print( "-----------------------------------")
|
||||
|
||||
# Routes for each panel
|
||||
routes = get_all_routes();
|
||||
lc_routes =lc_get_all_routes();
|
||||
|
||||
#File names
|
||||
controller_file = 'Targets/'+hl('PackageName')+'.pm'
|
||||
custom_controller_file = 'Targets/'+hl('PackageName')+'-Custom.pm'
|
||||
# Call it .new if one is already there (and may have been editted by the developer)
|
||||
if os.path.exists(custom_controller_file):
|
||||
custom_controller_file = custom_controller_file+'.new'
|
||||
layout_file = 'Targets/'+hl('PackageName').lower()+'.html.ep'
|
||||
partial_files = list()
|
||||
for panel in routes:
|
||||
partial_files.append('Targets/_'+hl('prefix')+"_"+panel+'.html.ep')
|
||||
print(partial_files)
|
||||
lex_file = 'Targets/'+hl('PackageName').lower()+'_en.lex'
|
||||
tablecontrols = get_table_control_data() #arrays of hashes used to drive rows in tables
|
||||
|
||||
#Generate controller file
|
||||
try:
|
||||
controller_template = PageTemplateFile("Templates/controller.pm.tem")
|
||||
dbentries = get_db_fields() #Params which correspond to Db fields
|
||||
try:
|
||||
controller_perl = controller_template.render(version=strVersion,
|
||||
tablecontrols=tablecontrols,
|
||||
dbentries=dbentries,
|
||||
**json5_dict,
|
||||
panels=routes,
|
||||
lcPackageName=json5_dict['PackageName'].lower()
|
||||
)
|
||||
with open(controller_file, 'w') as file:
|
||||
file.write(controller_perl)
|
||||
print(f"{controller_file} controller generated ok")
|
||||
except Exception as e:
|
||||
print(f"An chameleon controller render error occurred: {e}")
|
||||
except Exception as e:
|
||||
print(f"An chameleon controller template error occurred: {e}")
|
||||
|
||||
#Generate Custom controller file
|
||||
try:
|
||||
custom_controller_template = PageTemplateFile("Templates/custom.pm.tem")
|
||||
try:
|
||||
custom_controller_perl = custom_controller_template.render(version=strVersion,
|
||||
with open(controller_file, 'w') as file:
|
||||
file.write(controller_perl)
|
||||
print(f"{controller_file} controller generated ok")
|
||||
except Exception as e:
|
||||
print(f"A Chameleon controller render error occurred: {e}")
|
||||
except Exception as e:
|
||||
print(f"A Chameleon controller template error occurred: {e}")
|
||||
|
||||
#Generate Custom controller file
|
||||
try:
|
||||
custom_controller_template = PageTemplateFile("Templates/custom.pm.tem")
|
||||
try:
|
||||
custom_controller_perl = custom_controller_template.render(version=strVersion,
|
||||
panels=routes,
|
||||
tablecontrols=tablecontrols
|
||||
)
|
||||
# We must be careful to not overwrite the custom file if the developer has already written to it - TBD
|
||||
with open(custom_controller_file, 'w') as file:
|
||||
file.write(custom_controller_perl)
|
||||
print(f"{custom_controller_file} custom controller generated ok")
|
||||
except Exception as e:
|
||||
print(f"An chameleon custom controller render error occurred: {e}")
|
||||
except Exception as e:
|
||||
print(f"An chameleon custom controller template error occurred: {e}")
|
||||
# We must be careful to not overwrite the custom file if the developer has already written to it - TBD
|
||||
with open(custom_controller_file, 'w') as file:
|
||||
file.write(custom_controller_perl)
|
||||
print(f"{custom_controller_file} custom controller generated ok")
|
||||
except Exception as e:
|
||||
print(f"A Chameleon custom controller render error occurred: {e}")
|
||||
except Exception as e:
|
||||
print(f"A Chameleon custom controller template error occurred: {e}")
|
||||
|
||||
#generate Layout file
|
||||
layout_template = PageTemplateFile("Templates/layout.html.ep.tem")
|
||||
try:
|
||||
try:
|
||||
layout_mojo = layout_template.render(version=strVersion,**json5_dict,conditions=routes)
|
||||
with open(layout_file, 'w') as file:
|
||||
file.write(layout_mojo)
|
||||
print(f"{layout_file} mojo template layout file generated ok")
|
||||
except Exception as e:
|
||||
print(f"An chameleon render on layout file error occurred: {e}")
|
||||
except Exception as e:
|
||||
print(f"An chameleon template layout file error occurred: {e}")
|
||||
|
||||
#Generate a partial file for each of the entries in the html list
|
||||
#Pull in the template code for each of the input types
|
||||
#html_controls = json5_to_dict('Templates/html_controls.html.ep.tem')
|
||||
html_controls = parse_xml_to_dict('Templates/html_controls.html.ep.xml')
|
||||
i = 0
|
||||
for html in json5_html_list:
|
||||
# Generate a mojo template file, and then add in the controls
|
||||
# main file first
|
||||
try:
|
||||
partial_template = PageTemplateFile("Templates/partial.html.ep.tem")
|
||||
partial_mojo_context = {**json5_dict,**html}
|
||||
try:
|
||||
partial_mojo_template = partial_template.render(version=strVersion,**partial_mojo_context)
|
||||
with open( partial_files[i], 'w') as file:
|
||||
file.write(partial_mojo_template)
|
||||
print(f"{partial_files[i]} mojo template generated ok - phase 1")
|
||||
except Exception as e:
|
||||
print(f"An chameleon render error on partial file {html['route']} occurred: {e}")
|
||||
except Exception as e:
|
||||
print(f"An chameleon html {html['route']} error occurred: {e}")
|
||||
#generate Layout file
|
||||
layout_template = PageTemplateFile("Templates/layout.html.ep.tem")
|
||||
try:
|
||||
try:
|
||||
layout_mojo = layout_template.render(version=strVersion,**json5_dict,conditions=routes)
|
||||
with open(layout_file, 'w') as file:
|
||||
file.write(layout_mojo)
|
||||
print(f"{layout_file} mojo template layout file generated ok")
|
||||
except Exception as e:
|
||||
print(f"A Chameleon render on layout file error occurred: {e}")
|
||||
except Exception as e:
|
||||
print(f"A Chameleon template layout file error occurred: {e}")
|
||||
|
||||
#Generate a partial file for each of the entries in the html list
|
||||
#Pull in the template code for each of the input types
|
||||
#html_controls = json5_to_dict('Templates/html_controls.html.ep.tem')
|
||||
html_controls = parse_xml_to_dict('Templates/html_controls.html.ep.xml')
|
||||
i = 0
|
||||
for html in json5_html_list:
|
||||
# Generate a mojo template file, and then add in the controls
|
||||
# main file first
|
||||
try:
|
||||
partial_template = PageTemplateFile("Templates/partial.html.ep.tem")
|
||||
partial_mojo_context = {**json5_dict,**html}
|
||||
try:
|
||||
partial_mojo_template = partial_template.render(version=strVersion,**partial_mojo_context)
|
||||
with open( partial_files[i], 'w') as file:
|
||||
file.write(partial_mojo_template)
|
||||
print(f"{partial_files[i]} mojo template generated ok - phase 1")
|
||||
except Exception as e:
|
||||
print(f"A Chameleon render error on partial file {html['route']} occurred: {e}")
|
||||
except Exception as e:
|
||||
print(f"A Chameleon html {html['route']} error occurred: {e}")
|
||||
|
||||
#Now generate the controls from the rest of the entries in the dict.
|
||||
all_controls_html = "";
|
||||
prefix_is = hl('prefix')
|
||||
for html_control in html:
|
||||
inner_html = html[html_control]
|
||||
if isinstance(inner_html, dict):
|
||||
try:
|
||||
control_template = PageTemplate(html_controls[inner_html['Type']])
|
||||
try:
|
||||
control_html = control_template.render(version=strVersion,**inner_html,prefix=prefix_is)
|
||||
all_controls_html = all_controls_html + control_html
|
||||
except Exception as e:
|
||||
print(f"An chameleon render on partial file control {inner_html['Name']} error occurred: {e}")
|
||||
except Exception as e:
|
||||
print(f"An chameleon render on partial file control {inner_html['Name']} error occurred: {e}")
|
||||
else:
|
||||
#just a simple entry - name less numerics is type
|
||||
html_Type = ''.join(char for char in html_control if not char.isdigit())
|
||||
try:
|
||||
simple_control_template = PageTemplate(html_controls[html_Type])
|
||||
try:
|
||||
simple_control_html = simple_control_template.render(version=strVersion,value=inner_html,prefix=prefix_is)
|
||||
all_controls_html = all_controls_html + simple_control_html
|
||||
except Exception as e:
|
||||
print(f"An chameleon render on partial file control {html_control} error occurred: {e}")
|
||||
except Exception as e:
|
||||
print(f"An chameleon template partial file control {html_control} error occurred: {e}")
|
||||
|
||||
# Now insert it into the partial file in the correct place.
|
||||
# Read in the text file and split at "%# Inputs etc in here."
|
||||
with open(partial_files[i], 'r') as file:
|
||||
lines = file.readlines()
|
||||
index = next((i for i, line in enumerate(lines) if "%# Inputs etc in here." in line), len(lines))
|
||||
|
||||
# Insert the string at the specified index
|
||||
lines.insert(index+1, all_controls_html + '\n')
|
||||
|
||||
# Write the modified content back to the file
|
||||
with open(partial_files[i], 'w') as file:
|
||||
file.writelines(lines)
|
||||
print(f"Content modified and saved to {partial_files[i]}")
|
||||
i += 1
|
||||
|
||||
# Now generate the <name>.en file
|
||||
# Look through the generated files for the /l[\s|(]['|"](.*)['|"]\)/ strings.
|
||||
#Now generate the controls from the rest of the entries in the dict.
|
||||
all_controls_html = "";
|
||||
prefix_is = hl('prefix')
|
||||
for html_control in html:
|
||||
inner_html = html[html_control]
|
||||
if isinstance(inner_html, dict):
|
||||
try:
|
||||
control_template = PageTemplate(html_controls[inner_html['Type']])
|
||||
try:
|
||||
control_html = control_template.render(version=strVersion,**inner_html,prefix=prefix_is)
|
||||
all_controls_html = all_controls_html + control_html
|
||||
except Exception as e:
|
||||
print(f"A Chameleon render on partial file control {inner_html['Name']} error occurred: {e}")
|
||||
except Exception as e:
|
||||
print(f"A Chameleon render on partial file control {inner_html['Name']} error occurred: {e}")
|
||||
else:
|
||||
#just a simple entry - name less numerics is type
|
||||
html_Type = ''.join(char for char in html_control if not char.isdigit())
|
||||
try:
|
||||
simple_control_template = PageTemplate(html_controls[html_Type])
|
||||
try:
|
||||
simple_control_html = simple_control_template.render(version=strVersion,value=inner_html,prefix=prefix_is)
|
||||
all_controls_html = all_controls_html + simple_control_html
|
||||
except Exception as e:
|
||||
print(f"A Chameleon render on partial file control {html_control} error occurred: {e}")
|
||||
except Exception as e:
|
||||
print(f"A Chameleon template partial file control {html_control} error occurred: {e}")
|
||||
|
||||
# Now insert it into the partial file in the correct place.
|
||||
# Read in the text file and split at "%# Inputs etc in here."
|
||||
with open(partial_files[i], 'r') as file:
|
||||
lines = file.readlines()
|
||||
index = next((i for i, line in enumerate(lines) if "%# Inputs etc in here." in line), len(lines))
|
||||
|
||||
# Insert the string at the specified index
|
||||
lines.insert(index+1, all_controls_html + '\n')
|
||||
|
||||
# Write the modified content back to the file
|
||||
with open(partial_files[i], 'w') as file:
|
||||
file.writelines(lines)
|
||||
print(f"Content modified and saved to {partial_files[i]}")
|
||||
i += 1
|
||||
|
||||
# Now generate the <name>.en file
|
||||
# Look through the generated files for the /l[\s|(]['|"](.*)['|"]\)/ strings.
|
||||
|
||||
# create a combined list of all the files
|
||||
all_files = [controller_file,layout_file]+partial_files
|
||||
all_strings = []
|
||||
for filename in all_files:
|
||||
with open(filename, 'r') as file:
|
||||
file_content = file.read()
|
||||
# Define the regular expression pattern to match the strings you want to extract
|
||||
pattern = r"l[\s|(][\'|\"](.*)[\'|\"]\)"
|
||||
# Use re.findall to extract all occurrences of the pattern from the file content
|
||||
extracted_strings = re.findall(pattern, file_content)
|
||||
all_strings = all_strings + extracted_strings
|
||||
#Take out any duplicates
|
||||
all_strings = deduplicate_array(all_strings)
|
||||
# '<prefix>_english-message' => 'English Message',
|
||||
string_lib = [] #Array of dicts
|
||||
for lex_message in all_strings:
|
||||
# If has a prefix - leave it for left hand side but delete it for the right
|
||||
# If has no prefix - add one for left hand side but and leave it for the right
|
||||
# Map all spaces to "_" on left hand side
|
||||
# amd truncate it to max five words
|
||||
original_str = lex_message
|
||||
# Checkif it starts with the prefix (any case|)
|
||||
if lex_message.lower().startswith(hl('prefix').lower()):
|
||||
left_str = lex_message
|
||||
right_str = lex_message[len(hl('prefix'))+1:]
|
||||
# And take out any "_", map to " "
|
||||
else:
|
||||
left_str = hl('prefix')+"_"+lex_message
|
||||
right_str = lex_message
|
||||
right_str = right_str.replace("_"," ")
|
||||
print(f"Right:{right_str}")
|
||||
right_str = format_text(right_str)
|
||||
left_str = left_str.replace(" ","_")
|
||||
words = left_str.split('_')[:6]
|
||||
left_str = "_".join(words)
|
||||
next_lex_str = {"orig":original_str,"left":left_str,"right":right_str}
|
||||
string_lib.append(next_lex_str)
|
||||
#And write it to lex file
|
||||
# Now process them one by one into the lexical file
|
||||
lex_all = "";
|
||||
for lex_str in string_lib:
|
||||
lex_all += f"\'{lex_str['left']}\' => \'{lex_str['right']}\',\n"
|
||||
print(f"Writing {lex_file}")
|
||||
with open( lex_file, 'w') as file:
|
||||
file.write(lex_all)
|
||||
#and then play the strings back into the partials and the layout file
|
||||
print("..and feed the lex string names back into other files")
|
||||
for filename in all_files:
|
||||
with open(filename, 'r') as file:
|
||||
file_content = file.read()
|
||||
# Scan through
|
||||
for item in string_lib:
|
||||
original_str = item["orig"]
|
||||
left_str = item["left"]
|
||||
right_str = item["right"]
|
||||
# Replace all occurrences of original string with left string in 'contents'
|
||||
file_content = file_content.replace("l('"+original_str+"')", "l('"+left_str+"')")
|
||||
# and write it back
|
||||
with open(filename, 'w') as file:
|
||||
file.write(file_content)
|
||||
print(f"Write out modified:{filename}")
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
# create a combined list of all the files
|
||||
all_files = [controller_file,layout_file]+partial_files
|
||||
all_strings = []
|
||||
for filename in all_files:
|
||||
with open(filename, 'r') as file:
|
||||
file_content = file.read()
|
||||
# Define the regular expression pattern to match the strings you want to extract
|
||||
pattern = r"l[\s|(][\'|\"](.*)[\'|\"]\)"
|
||||
# Use re.findall to extract all occurrences of the pattern from the file content
|
||||
extracted_strings = re.findall(pattern, file_content)
|
||||
all_strings = all_strings + extracted_strings
|
||||
#Take out any duplicates
|
||||
all_strings = deduplicate_array(all_strings)
|
||||
# '<prefix>_english-message' => 'English Message',
|
||||
string_lib = [] #Array of dicts
|
||||
for lex_message in all_strings:
|
||||
# If has a prefix - leave it for left hand side but delete it for the right
|
||||
# If has no prefix - add one for left hand side but and leave it for the right
|
||||
# Map all spaces to "_" on left hand side
|
||||
# amd truncate it to max five words
|
||||
original_str = lex_message
|
||||
# Checkif it starts with the prefix (any case|)
|
||||
if lex_message.lower().startswith(hl('prefix').lower()):
|
||||
left_str = lex_message
|
||||
right_str = lex_message[len(hl('prefix'))+1:]
|
||||
# And take out any "_", map to " "
|
||||
else:
|
||||
left_str = hl('prefix')+"_"+lex_message
|
||||
right_str = lex_message
|
||||
right_str = right_str.replace("_"," ")
|
||||
#print(f"Right:{right_str}")
|
||||
right_str = format_text(right_str)
|
||||
left_str = left_str.replace(" ","_")
|
||||
words = left_str.split('_')[:6]
|
||||
left_str = "_".join(words)
|
||||
next_lex_str = {"orig":original_str,"left":left_str,"right":right_str}
|
||||
string_lib.append(next_lex_str)
|
||||
#And write it to lex file
|
||||
# Now process them one by one into the lexical file
|
||||
lex_all = "";
|
||||
for lex_str in string_lib:
|
||||
lex_all += f"\'{lex_str['left']}\' => \'{lex_str['right']}\',\n"
|
||||
print(f"Writing {lex_file}")
|
||||
with open( lex_file, 'w') as file:
|
||||
file.write(lex_all)
|
||||
#and then play the strings back into the partials and the layout file
|
||||
print("..and feed the lex string names back into other files")
|
||||
for filename in all_files:
|
||||
with open(filename, 'r') as file:
|
||||
file_content = file.read()
|
||||
# Scan through
|
||||
for item in string_lib:
|
||||
original_str = item["orig"]
|
||||
left_str = item["left"]
|
||||
right_str = item["right"]
|
||||
# Replace all occurrences of original string with left string in 'contents'
|
||||
file_content = file_content.replace("l('"+original_str+"')", "l('"+left_str+"')")
|
||||
# and write it back
|
||||
with open(filename, 'w') as file:
|
||||
file.write(file_content)
|
||||
print(f"Write out modified:{filename}")
|
||||
|
||||
# Now generate all the translated lex files from a list of the languages and codes
|
||||
home_dir = os.path.dirname(json_filename)
|
||||
languages_path =f"{home_dir}/languages.json"
|
||||
with open(languages_path,'r') as file:
|
||||
languages_str = file.read()
|
||||
lang_dict = json.loads(languages_str)
|
||||
with open(lex_file,'r') as file:
|
||||
lex_str = file.read()
|
||||
eng_lex_dict = convert_lex_to_dict(lex_str)
|
||||
for lang_item in lang_dict:
|
||||
print(f"Translating from english lex file to {lang_item['language']}")
|
||||
code = lang_item["code"]
|
||||
translated_dict = []
|
||||
for lex_item in eng_lex_dict:
|
||||
translated_text = get_translation(lex_item["text"], lang_item["language"])
|
||||
translated_dict.append({"id": lex_item["id"], "text": translated_text})
|
||||
print(f"Writing out lex file for {lang_item['code']}")
|
||||
translated_lex_file = f"Targets/{hl('PackageName').lower()}_{code}.lex"
|
||||
with open(translated_lex_file, "w") as file:
|
||||
for item in translated_dict:
|
||||
file.write(f"'"+item['id']+"' => '"+item['text']+"'\n")
|
||||
#print(f"{item['id']} => {item['text']}\n")
|
||||
|
||||
|
Reference in New Issue
Block a user