# First Attempt
- Without any preprocessing
- Parameters like these
```bash
"task": "MS",
"lstm_layer": 1,
"dropout": 0.05,
"train_epochs": 1,
"batch_size": 64,
"patience": 3,
"lr": 1e-4,
"lr_adjust": "type1",
```
### Result
Score: -55.30953
---
# Problems
- Slow Training:
It take about 6 hours to train all the turbine, and the evaluation set does not really reflect the final score
- Low Accuracy:
Because the first attempt is for speed only, we need to improve the accuracy
---
# 70% Faster … sort of
So we split the training work into 4 partition on 4 different Colab machine (manually). So it’s roughly 70% faster now considering the manual operation.
---
# Grid search
we use grid search method to try all the combination on turbine 0 ~ 10
```python
lstm_layer = [2,3,4]
drop_out = [0,0.05,0.1,0.3]
batch_size=[16,32,64]
lr = [1e-3, 1e-4, 1e-5]
```
And found that baseline code default yield the best result
### Result
-42.3428 (Big improvement)
---
# GRU → LSTM
- The mechnism of LSTM and GRU is similar (GRU is more like a subcategory of LSTM)
- But GRU has two gates and LSTM has three gates to decide the in-n-out of the information
- Since LSTM has more gate, we assume that LSTM will be more accurate
- We test it without parameter chage. The accuracy are around the same, but spend more time on training for LSTM
- GRU is more efficient
---
# Increase Train size
Train: Test: Validation = 214:15:31 → 235:10:0
---
# Pre-processing Experiments
- Fill all the unqualified data with previous and next qualified data average of the same turbine
Result: -42.18417 (New Best)
- Fill with average of average of ( data from last normal day and the next normal day at the same time stamp
Result: -42.37965
---
# Add more features to Training
- We tried to take Time stamp and Day into training with the second processing method
Result: -42.3459
---
# What we have learned
- Splitting the training data make the speed of traing faster than before. We now know how to deal with the dataset with large scale.
- An appropriate Model is better than a cool Model
- Pre-processing Experiments is very very important since that "garbage in garbage out".
{"metaMigratedAt":"2023-06-17T03:01:56.946Z","metaMigratedFrom":"Content","title":"First Attempt","breaks":true,"contributors":"[{\"id\":\"200300e0-32e2-4edc-968b-2d8433bd9742\",\"add\":301,\"del\":27},{\"id\":\"14193e02-6de4-4911-af57-a5c10383e171\",\"add\":1952,\"del\":38}]"}