feat: enhance logging and memory management across modules

This commit is contained in:
2025-08-13 23:09:05 +02:00
parent 47e52d130c
commit 9601c5e44d
48 changed files with 1004 additions and 773 deletions
+63 -25
View File
@@ -2,51 +2,89 @@
import argparse
import logging
import sys
import os
import traceback
from pathlib import Path
from core.config import setup_config
from core.config import setup_config, PipelineConfig
from processing.ner.ner_data_builder import NERDataBuilder
from processing.ner.ner_engineering import NEREngineering
from processing.ner.ner_name_model import NERNameModel
def train(config_path=None, env="development"):
def feature(config: PipelineConfig):
"""Apply feature engineering to create position-independent NER dataset."""
NEREngineering(config).compute()
def build(config: PipelineConfig):
"""Build NER dataset using NERDataBuilder."""
NERDataBuilder(config).build()
def train(config: PipelineConfig):
"""Train the NER model."""
try:
config = setup_config(config_path=config_path, env=env)
trainer = NERNameModel(config)
builder = NERDataBuilder(config)
trainer = NERNameModel(config)
data_path = Path(config.paths.data_dir) / config.data.output_files["ner_data"]
if not data_path.exists():
builder.build()
data_path = Path(config.paths.data_dir) / config.data.output_files["ner_data"]
if not data_path.exists():
logging.info("NER data not found. Building dataset first...")
build(config)
trainer.create_blank_model("fr")
data = trainer.load_data(str(data_path))
trainer.create_blank_model("fr")
data = trainer.load_data(str(data_path))
split_idx = int(len(data) * 0.8)
train_data, eval_data = data[:split_idx], data[split_idx:]
split_idx = int(len(data) * 0.9)
train_data, eval_data = data[:split_idx], data[split_idx:]
logging.info(f"Training with {len(train_data)} examples, evaluating on {len(eval_data)}")
trainer.train(train_data, epochs=1, batch_size=config.processing.batch_size, dropout_rate=0.3)
trainer.evaluate(eval_data)
logging.info(f"Training with {len(train_data)} examples, evaluating on {len(eval_data)}")
trainer.train(
data=train_data, epochs=1, batch_size=config.processing.batch_size, dropout_rate=0.3
)
trainer.evaluate(eval_data)
model_path = trainer.save()
logging.info(f"Model saved to: {model_path}")
return 0
model_path = trainer.save()
logging.info(f"Model saved to: {model_path}")
except Exception as e:
logging.error(f"NER Training failed: {e}", exc_info=True)
return 1
def run_pipeline(config: PipelineConfig, reset: bool = False):
# Step 1: Feature engineering
if not reset and os.path.exists(config.paths.data_dir / config.data.output_files["engineered"]):
logging.info("Step 1: Feature engineering already done.")
else:
logging.info("Step 1: Running feature engineering")
feature(config)
# Step 2: Build dataset
if not reset and os.path.exists(config.paths.data_dir / config.data.output_files["ner_data"]):
logging.info("Step 2: NER dataset already built.")
else:
logging.info("Step 2: Building NER dataset")
build(config)
# Step 3: Train model
logging.info("Step 3: Training NER Model")
train(config)
return 0
def main():
parser = argparse.ArgumentParser(description="Train NER model for DRC names")
parser = argparse.ArgumentParser(description="NER model management for DRC names")
parser.add_argument("--config", type=str, help="Path to configuration file")
parser.add_argument("--env", type=str, default="development", help="Environment name")
parser.add_argument("--reset", action="store_true", help="Reset all steps")
args = parser.parse_args()
sys.exit(train(config_path=args.config, env=args.env))
try:
config = setup_config(config_path=args.config, env=args.env)
return run_pipeline(config, args.reset)
except Exception as e:
print(f"Pipeline failed: {e}")
traceback.print_exc()
return 1
if __name__ == "__main__":
main()
sys.exit(main())