You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@madlib.apache.org by "Frank McQuillan (Jira)" <ji...@apache.org> on 2020/11/20 18:06:00 UTC
[jira] [Closed] (MADLIB-1454) DL - Write best so far to console for
autoML methods
[ https://issues.apache.org/jira/browse/MADLIB-1454?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Frank McQuillan closed MADLIB-1454.
-----------------------------------
Resolution: Fixed
done as part of
https://github.com/apache/madlib/pull/519
> DL - Write best so far to console for autoML methods
> -----------------------------------------------------
>
> Key: MADLIB-1454
> URL: https://issues.apache.org/jira/browse/MADLIB-1454
> Project: Apache MADlib
> Issue Type: Improvement
> Components: Deep Learning
> Reporter: Frank McQuillan
> Assignee: Advitya Gemawat
> Priority: Minor
> Fix For: v1.18.0
>
>
> For Hyperband, write the "best so far" to the console so that user knows how things are progressing.
> Note need to keep track of global best, it might not be the one from the last iteration.
> Change console output from:
> {code}
> INFO: *** Diagonally evaluating 9 configs under bracket=2 & round=0 with 1 iterations ***
> CONTEXT: PL/Python function "madlib_keras_automl"
> INFO:
> Time for training in iteration 1: 9.76507210732 sec
> DETAIL:
> Training set after iteration 1:
> mst_key=2: metric=0.683333337307, loss=0.626947939396
> mst_key=8: metric=0.683333337307, loss=0.556752383709
> mst_key=3: metric=0.683333337307, loss=0.604624867439
> mst_key=6: metric=0.324999988079, loss=1.01775479317
> mst_key=1: metric=0.691666662693, loss=0.918690085411
> mst_key=7: metric=0.324999988079, loss=1.09102141857
> mst_key=9: metric=0.683333337307, loss=0.615454554558
> mst_key=4: metric=0.774999976158, loss=0.571036159992
> mst_key=5: metric=0.324999988079, loss=1.10194396973
> Validation set after iteration 1:
> mst_key=2: metric=0.600000023842, loss=0.67598927021
> mst_key=8: metric=0.600000023842, loss=0.62441021204
> mst_key=3: metric=0.600000023842, loss=0.669852972031
> mst_key=6: metric=0.366666674614, loss=0.984160840511
> mst_key=1: metric=0.600000023842, loss=0.923334658146
> mst_key=7: metric=0.366666674614, loss=1.07771503925
> mst_key=9: metric=0.600000023842, loss=0.699421286583
> mst_key=4: metric=0.866666674614, loss=0.607381045818
> mst_key=5: metric=0.366666674614, loss=1.09954810143
> CONTEXT: PL/Python function "madlib_keras_automl"
> INFO: *** Diagonally evaluating 3 configs under bracket=2 & round=1, 3 configs under bracket=1 & round=0 with 3 iterations ***
> CONTEXT: PL/Python function "madlib_keras_automl"
> INFO:
> Time for training in iteration 1: 4.84015893936 sec
> DETAIL:
> Training set after iteration 1:
> mst_key=8: metric=0.925000011921, loss=0.353324443102
> mst_key=4: metric=0.949999988079, loss=0.424594521523
> mst_key=11: metric=0.675000011921, loss=0.846702694893
> mst_key=3: metric=0.808333337307, loss=0.382121056318
> mst_key=12: metric=0.916666686535, loss=0.384196609259
> mst_key=10: metric=0.683333337307, loss=0.701473772526
> Validation set after iteration 1:
> mst_key=8: metric=0.933333337307, loss=0.42084941268
> mst_key=4: metric=0.933333337307, loss=0.476406633854
> mst_key=11: metric=0.600000023842, loss=0.854079544544
> mst_key=3: metric=0.899999976158, loss=0.417265832424
> mst_key=12: metric=0.899999976158, loss=0.450416505337
> mst_key=10: metric=0.600000023842, loss=0.728042304516
> CONTEXT: PL/Python function "madlib_keras_automl"
> INFO:
> Time for training in iteration 2: 4.80181288719 sec
> DETAIL:
> Training set after iteration 2:
> mst_key=8: metric=0.941666662693, loss=0.286089539528
> mst_key=4: metric=0.925000011921, loss=0.373028248549
> mst_key=11: metric=0.683333337307, loss=0.609232187271
> mst_key=3: metric=0.833333313465, loss=0.291878581047
> mst_key=12: metric=0.908333361149, loss=0.300016224384
> mst_key=10: metric=0.983333349228, loss=0.382896214724
> Validation set after iteration 2:
> mst_key=8: metric=0.933333337307, loss=0.338641613722
> mst_key=4: metric=1.0, loss=0.436057478189
> mst_key=11: metric=0.600000023842, loss=0.658753097057
> mst_key=3: metric=0.766666650772, loss=0.339546382427
> mst_key=12: metric=0.933333337307, loss=0.341486483812
> mst_key=10: metric=1.0, loss=0.442664504051
> CONTEXT: PL/Python function "madlib_keras_automl"
> INFO:
> Time for training in iteration 3: 5.17401909828 sec
> DETAIL:
> Training set after iteration 3:
> mst_key=8: metric=0.966666638851, loss=0.196135208011
> mst_key=4: metric=0.958333313465, loss=0.243382230401
> mst_key=11: metric=0.941666662693, loss=0.395315974951
> mst_key=3: metric=0.966666638851, loss=0.171766787767
> mst_key=12: metric=0.866666674614, loss=0.283820331097
> mst_key=10: metric=0.833333313465, loss=0.313775897026
> Validation set after iteration 3:
> mst_key=8: metric=0.966666638851, loss=0.214255988598
> mst_key=4: metric=1.0, loss=0.268849998713
> mst_key=11: metric=0.899999976158, loss=0.45996800065
> mst_key=3: metric=1.0, loss=0.157373458147
> mst_key=12: metric=0.800000011921, loss=0.340971261263
> mst_key=10: metric=0.766666650772, loss=0.365937292576
> CONTEXT: PL/Python function "madlib_keras_automl"
> {code}
> to
> {code}
> INFO: *** Diagonally evaluating 9 configs under bracket=2 & round=0 with 1 iterations ***
> CONTEXT: PL/Python function "madlib_keras_automl"
> INFO:
> Time for training in iteration 1: 9.76507210732 sec
> DETAIL:
> Training set after iteration 1:
> mst_key=2: metric=0.683333337307, loss=0.626947939396
> mst_key=8: metric=0.683333337307, loss=0.556752383709
> mst_key=3: metric=0.683333337307, loss=0.604624867439
> mst_key=6: metric=0.324999988079, loss=1.01775479317
> mst_key=1: metric=0.691666662693, loss=0.918690085411
> mst_key=7: metric=0.324999988079, loss=1.09102141857
> mst_key=9: metric=0.683333337307, loss=0.615454554558
> mst_key=4: metric=0.774999976158, loss=0.571036159992
> mst_key=5: metric=0.324999988079, loss=1.1019439697
> Validation set after iteration 1:
> mst_key=2: metric=0.600000023842, loss=0.67598927021
> mst_key=8: metric=0.600000023842, loss=0.62441021204
> mst_key=3: metric=0.600000023842, loss=0.669852972031
> mst_key=6: metric=0.366666674614, loss=0.984160840511
> mst_key=1: metric=0.600000023842, loss=0.923334658146
> mst_key=7: metric=0.366666674614, loss=1.07771503925
> mst_key=9: metric=0.600000023842, loss=0.699421286583
> mst_key=4: metric=0.866666674614, loss=0.607381045818
> mst_key=5: metric=0.366666674614, loss=1.09954810143
> INFO:
> Best training metric so far:
> mst_key=4: metric=0.774999976158, loss=0.571036159992
> Best validation metric so far:
> mst_key=4: metric=0.866666674614, loss=0.607381045818
> CONTEXT: PL/Python function "madlib_keras_automl"
> INFO: *** Diagonally evaluating 3 configs under bracket=2 & round=1, 3 configs under bracket=1 & round=0 with 3 iterations ***
> CONTEXT: PL/Python function "madlib_keras_automl"
> INFO:
> Time for training in iteration 1: 4.84015893936 sec
> DETAIL:
> Training set after iteration 1:
> mst_key=8: metric=0.925000011921, loss=0.353324443102
> mst_key=4: metric=0.949999988079, loss=0.424594521523
> mst_key=11: metric=0.675000011921, loss=0.846702694893
> mst_key=3: metric=0.808333337307, loss=0.382121056318
> mst_key=12: metric=0.916666686535, loss=0.384196609259
> mst_key=10: metric=0.683333337307, loss=0.701473772526
> Validation set after iteration 1:
> mst_key=8: metric=0.933333337307, loss=0.42084941268
> mst_key=4: metric=0.933333337307, loss=0.476406633854
> mst_key=11: metric=0.600000023842, loss=0.854079544544
> mst_key=3: metric=0.899999976158, loss=0.417265832424
> mst_key=12: metric=0.899999976158, loss=0.450416505337
> mst_key=10: metric=0.600000023842, loss=0.728042304516
> CONTEXT: PL/Python function "madlib_keras_automl"
> INFO:
> Time for training in iteration 2: 4.80181288719 sec
> DETAIL:
> Training set after iteration 2:
> mst_key=8: metric=0.941666662693, loss=0.286089539528
> mst_key=4: metric=0.925000011921, loss=0.373028248549
> mst_key=11: metric=0.683333337307, loss=0.609232187271
> mst_key=3: metric=0.833333313465, loss=0.291878581047
> mst_key=12: metric=0.908333361149, loss=0.300016224384
> mst_key=10: metric=0.983333349228, loss=0.382896214724
> Validation set after iteration 2:
> mst_key=8: metric=0.933333337307, loss=0.338641613722
> mst_key=4: metric=1.0, loss=0.436057478189
> mst_key=11: metric=0.600000023842, loss=0.658753097057
> mst_key=3: metric=0.766666650772, loss=0.339546382427
> mst_key=12: metric=0.933333337307, loss=0.341486483812
> mst_key=10: metric=1.0, loss=0.442664504051
> CONTEXT: PL/Python function "madlib_keras_automl"
> INFO:
> Time for training in iteration 3: 5.17401909828 sec
> DETAIL:
> Training set after iteration 3:
> mst_key=8: metric=0.966666638851, loss=0.196135208011
> mst_key=4: metric=0.958333313465, loss=0.243382230401
> mst_key=11: metric=0.941666662693, loss=0.395315974951
> mst_key=3: metric=0.966666638851, loss=0.171766787767
> mst_key=12: metric=0.866666674614, loss=0.283820331097
> mst_key=10: metric=0.833333313465, loss=0.313775897026
> Validation set after iteration 3:
> mst_key=8: metric=0.966666638851, loss=0.214255988598
> mst_key=4: metric=1.0, loss=0.268849998713
> mst_key=11: metric=0.899999976158, loss=0.45996800065
> mst_key=3: metric=1.0, loss=0.157373458147
> mst_key=12: metric=0.800000011921, loss=0.340971261263
> mst_key=10: metric=0.766666650772, loss=0.365937292576
> INFO:
> Best training metric so far:
> mst_key=8: metric=0.966666638851, loss=0.196135208011
> Best validation metric so far:
> mst_key=8: metric=0.966666638851, loss=0.214255988598
> CONTEXT: PL/Python function "madlib_keras_automl"
> {code}
--
This message was sent by Atlassian Jira
(v8.3.4#803005)