Unterschiede
Hier werden die Unterschiede zwischen zwei Versionen angezeigt.
| Nächste Überarbeitung | Vorhergehende Überarbeitung | ||
| de:modul:m245:learningunits:lu02:loesungen:l02 [2026/01/05 13:26] – angelegt vdemir | de:modul:m245:learningunits:lu02:loesungen:l02 [2026/04/08 08:43] (aktuell) – [Modellvergleich] vdemir | ||
|---|---|---|---|
| Zeile 3: | Zeile 3: | ||
| ===== Voraussetzung ===== | ===== Voraussetzung ===== | ||
| - | | + | <code python> |
| - | + | pip install pandas scikit-learn joblib | |
| + | </ | ||
| ===== Python-Skript: | ===== Python-Skript: | ||
| + | <code python> | ||
| import pandas as pd | import pandas as pd | ||
| from sklearn.model_selection import train_test_split | from sklearn.model_selection import train_test_split | ||
| Zeile 15: | Zeile 17: | ||
| from sklearn.metrics import accuracy_score, | from sklearn.metrics import accuracy_score, | ||
| import joblib | import joblib | ||
| - | + | ||
| - | # ----------------------------- | + | # ----------------------------- |
| # Daten laden | # Daten laden | ||
| # ----------------------------- | # ----------------------------- | ||
| data = pd.read_csv(" | data = pd.read_csv(" | ||
| - | |||
| X = data.drop(" | X = data.drop(" | ||
| y = data[" | y = data[" | ||
| Zeile 27: | Zeile 28: | ||
| # Train / Test Split | # Train / Test Split | ||
| # ----------------------------- | # ----------------------------- | ||
| - | X_train, X_test, y_train, y_test = train_test_split( | + | X_train, X_test, y_train, y_test = train_test_split( X, y, test_size=0.2, |
| - | | + | |
| - | ) | + | |
| # ----------------------------- | # ----------------------------- | ||
| Zeile 41: | Zeile 40: | ||
| log_reg_pipeline.fit(X_train, | log_reg_pipeline.fit(X_train, | ||
| y_pred_lr = log_reg_pipeline.predict(X_test) | y_pred_lr = log_reg_pipeline.predict(X_test) | ||
| + | # | ||
| print(" | print(" | ||
| print(" | print(" | ||
| Zeile 77: | Zeile 76: | ||
| print(" | print(" | ||
| + | </ | ||
| + | | ||
| + | ===== Modellvergleich ===== | ||
| + | ^ Kriterium ^ Logistische Regression ^ Decision Tree ^ | ||
| + | | Interpretierbarkeit | hoch | mittel | | ||
| + | | Overfitting-Gefahr | gering | hoch | | ||
| + | | Skalierung | nötig ja | nein | | ||
| + | | Didaktisch | sinnvoll sehr | ja | | ||
| + | ===== Fazit ===== | ||
| + | * Bei kleinen, sauberen Datensätzen ist die Logistische Regression meist stabiler. | ||
| + | * Decision Trees sind anschaulich, | ||
| + | ---- | ||
| + | [[https:// | ||
| + | | ||