feat: enhance logging and memory management across modules
This commit is contained in:
@@ -50,14 +50,18 @@ class LightGBMModel(TraditionalModel):
|
||||
self.vectorizers[feature_key] = CountVectorizer(
|
||||
analyzer="char", ngram_range=(2, 3), max_features=50
|
||||
)
|
||||
char_features = self.vectorizers[feature_key].fit_transform(
|
||||
column.fillna("").astype(str)
|
||||
).toarray()
|
||||
char_features = (
|
||||
self.vectorizers[feature_key]
|
||||
.fit_transform(column.fillna("").astype(str))
|
||||
.toarray()
|
||||
)
|
||||
else:
|
||||
# Subsequent times - use existing vectorizer
|
||||
char_features = self.vectorizers[feature_key].transform(
|
||||
column.fillna("").astype(str)
|
||||
).toarray()
|
||||
char_features = (
|
||||
self.vectorizers[feature_key]
|
||||
.transform(column.fillna("").astype(str))
|
||||
.toarray()
|
||||
)
|
||||
|
||||
features.append(char_features)
|
||||
else:
|
||||
|
||||
@@ -20,9 +20,7 @@ class LogisticRegressionModel(TraditionalModel):
|
||||
)
|
||||
|
||||
classifier = LogisticRegression(
|
||||
max_iter=params.get("max_iter", 1000),
|
||||
random_state=self.config.random_seed,
|
||||
verbose=2
|
||||
max_iter=params.get("max_iter", 1000), random_state=self.config.random_seed, verbose=2
|
||||
)
|
||||
|
||||
return Pipeline([("vectorizer", vectorizer), ("classifier", classifier)])
|
||||
|
||||
@@ -18,7 +18,7 @@ class RandomForestModel(TraditionalModel):
|
||||
n_estimators=params.get("n_estimators", 100),
|
||||
max_depth=params.get("max_depth", None),
|
||||
random_state=self.config.random_seed,
|
||||
verbose=2
|
||||
verbose=2,
|
||||
)
|
||||
|
||||
def prepare_features(self, X: pd.DataFrame) -> np.ndarray:
|
||||
|
||||
@@ -25,7 +25,7 @@ class SVMModel(TraditionalModel):
|
||||
gamma=params.get("gamma", "scale"),
|
||||
probability=True, # Enable probability prediction
|
||||
random_state=self.config.random_seed,
|
||||
verbose=2
|
||||
verbose=2,
|
||||
)
|
||||
|
||||
return Pipeline([("vectorizer", vectorizer), ("classifier", classifier)])
|
||||
|
||||
@@ -28,7 +28,7 @@ class XGBoostModel(TraditionalModel):
|
||||
colsample_bytree=params.get("colsample_bytree", 0.8),
|
||||
random_state=self.config.random_seed,
|
||||
eval_metric="logloss",
|
||||
verbosity=2
|
||||
verbosity=2,
|
||||
)
|
||||
|
||||
def prepare_features(self, X: pd.DataFrame) -> np.ndarray:
|
||||
@@ -50,14 +50,18 @@ class XGBoostModel(TraditionalModel):
|
||||
self.vectorizers[feature_key] = CountVectorizer(
|
||||
analyzer="char", ngram_range=(2, 3), max_features=100
|
||||
)
|
||||
char_features = self.vectorizers[feature_key].fit_transform(
|
||||
column.fillna("").astype(str)
|
||||
).toarray()
|
||||
char_features = (
|
||||
self.vectorizers[feature_key]
|
||||
.fit_transform(column.fillna("").astype(str))
|
||||
.toarray()
|
||||
)
|
||||
else:
|
||||
# Subsequent times - use existing vectorizer
|
||||
char_features = self.vectorizers[feature_key].transform(
|
||||
column.fillna("").astype(str)
|
||||
).toarray()
|
||||
char_features = (
|
||||
self.vectorizers[feature_key]
|
||||
.transform(column.fillna("").astype(str))
|
||||
.toarray()
|
||||
)
|
||||
|
||||
features.append(char_features)
|
||||
else:
|
||||
|
||||
Reference in New Issue
Block a user