Skip to content

Commit e78971e

Browse files
artidoroshauheen
authored andcommitted
substituted winequality dataset with machine generated dataset (#937)
1 parent 350f77f commit e78971e

File tree

98 files changed

+7778
-118177
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

98 files changed

+7778
-118177
lines changed

build.proj

Lines changed: 1 addition & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -75,11 +75,7 @@
7575
Targets="Pack" />
7676
</Target>
7777

78-
<ItemGroup>
79-
<TestFile Include="$(MSBuildThisFileDirectory)/test/data/external/winequality-white.csv"
80-
Url="https://archive.ics.uci.edu/ml/machine-learning-databases/wine-quality/winequality-white.csv"
81-
DestinationFile="$(MSBuildThisFileDirectory)test/data/external/winequality-white.csv" />
82-
78+
<ItemGroup>
8379
<TestFile Condition="'$(IncludeBenchmarkData)' == 'true'" Include="$(MSBuildThisFileDirectory)/test/data/external/WikiDetoxAnnotated160kRows.tsv"
8480
Url="https://aka.ms/tlc-resources/benchmarks/WikiDetoxAnnotated160kRows.tsv"
8581
DestinationFile="$(MSBuildThisFileDirectory)test/data/external/WikiDetoxAnnotated160kRows.tsv" />

test/BaselineOutput/SingleDebug/LightGBMR/LightGBMReg-CV-wine-out.txt renamed to test/BaselineOutput/SingleDebug/LightGBMR/LightGBMReg-CV-generatedRegressionDataset-out.txt

Lines changed: 17 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -7,24 +7,24 @@ Not adding a normalizer.
77
Auto-tuning parameters: UseCat = False
88
LightGBM objective=regression
99
Not training a calibrator because it is not needed.
10-
L1(avg): 0.524348
11-
L2(avg): 0.466735
12-
RMS(avg): 0.683180
13-
Loss-fn(avg): 0.466735
14-
R Squared: 0.400415
15-
L1(avg): 0.517508
16-
L2(avg): 0.458039
17-
RMS(avg): 0.676786
18-
Loss-fn(avg): 0.458039
19-
R Squared: 0.420159
10+
L1(avg): 27.477977
11+
L2(avg): 1,428.594095
12+
RMS(avg): 37.796747
13+
Loss-fn(avg): 1,428.594094
14+
R Squared: 0.920504
15+
L1(avg): 26.801569
16+
L2(avg): 1,413.398603
17+
RMS(avg): 37.595194
18+
Loss-fn(avg): 1,413.398596
19+
R Squared: 0.923322
2020

2121
OVERALL RESULTS
2222
---------------------------------------
23-
L1(avg): 0.520928 (0.0034)
24-
L2(avg): 0.462387 (0.0043)
25-
RMS(avg): 0.679983 (0.0032)
26-
Loss-fn(avg): 0.462387 (0.0043)
27-
R Squared: 0.410287 (0.0099)
23+
L1(avg): 27.139773 (0.3382)
24+
L2(avg): 1,420.996349 (7.5977)
25+
RMS(avg): 37.695971 (0.1008)
26+
Loss-fn(avg): 1,420.996345 (7.5977)
27+
R Squared: 0.921913 (0.0014)
2828

2929
---------------------------------------
3030
Physical memory usage(MB): %Number%
@@ -35,10 +35,10 @@ Virtual memory usage(MB): %Number%
3535
[1] 'Loading data for LightGBM' started.
3636
[1] 'Loading data for LightGBM' finished in %Time%.
3737
[2] 'Training with LightGBM' started.
38-
[2] (%Time%) Iteration: 50 Training-l2: 0.189697165394939
38+
[2] (%Time%) Iteration: 50 Training-l2: 37.107605006517
3939
[2] 'Training with LightGBM' finished in %Time%.
4040
[3] 'Loading data for LightGBM #2' started.
4141
[3] 'Loading data for LightGBM #2' finished in %Time%.
4242
[4] 'Training with LightGBM #2' started.
43-
[4] (%Time%) Iteration: 50 Training-l2: 0.204982247876212
43+
[4] (%Time%) Iteration: 50 Training-l2: 27.7037679135951
4444
[4] 'Training with LightGBM #2' finished in %Time%.
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
11
LightGBMR
22
L1(avg) L2(avg) RMS(avg) Loss-fn(avg) R Squared /iter /lr /nl /mil /booster /v /nt Learner Name Train Dataset Test Dataset Results File Run Time Physical Memory Virtual Memory Command Line Settings
3-
0.520928 0.462387 0.679983 0.462387 0.410287 50 0.2 20 10 gbdt{l2=0.2 l1=0.2} + 1 LightGBMR %Data% %Output% 99 0 0 maml.exe CV tr=LightGBMR{nt=1 iter=50 v=+ booster=gbdt{l1=0.2 l2=0.2} lr=0.2 mil=10 nl=20} threads=- dout=%Output% loader=Text{col=Label:R4:11 col=Features:R4:0-10 sep=; header+} data=%Data% seed=1 /iter:50;/lr:0.2;/nl:20;/mil:10;/booster:gbdt{l2=0.2 l1=0.2};/v:+;/nt:1
3+
27.13977 1420.996 37.69597 1420.996 0.921913 50 0.2 20 10 gbdt{l2=0.2 l1=0.2} + 1 LightGBMR %Data% %Output% 99 0 0 maml.exe CV tr=LightGBMR{nt=1 iter=50 v=+ booster=gbdt{l1=0.2 l2=0.2} lr=0.2 mil=10 nl=20} threads=- dout=%Output% loader=Text{col=Label:R4:11 col=Features:R4:0-10 sep=; header+} data=%Data% seed=1 /iter:50;/lr:0.2;/nl:20;/mil:10;/booster:gbdt{l2=0.2 l1=0.2};/v:+;/nt:1
44

test/BaselineOutput/SingleDebug/LightGBMR/LightGBMReg-CV-generatedRegressionDataset.txt

Lines changed: 501 additions & 0 deletions
Large diffs are not rendered by default.

test/BaselineOutput/SingleDebug/LightGBMR/LightGBMReg-CV-wine.txt

Lines changed: 0 additions & 4899 deletions
This file was deleted.

test/BaselineOutput/SingleRelease/LightGBMR/LightGBMReg-TrainTest-wine-out.txt renamed to test/BaselineOutput/SingleDebug/LightGBMR/LightGBMReg-TrainTest-generatedRegressionDataset-out.txt

Lines changed: 11 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -3,19 +3,19 @@ Not adding a normalizer.
33
Auto-tuning parameters: UseCat = False
44
LightGBM objective=regression
55
Not training a calibrator because it is not needed.
6-
L1(avg): 0.402080
7-
L2(avg): 0.272274
8-
RMS(avg): 0.521799
9-
Loss-fn(avg): 0.272274
10-
R Squared: 0.652798
6+
L1(avg): 3.472291
7+
L2(avg): 26.064428
8+
RMS(avg): 5.105333
9+
Loss-fn(avg): 26.064428
10+
R Squared: 0.998571
1111

1212
OVERALL RESULTS
1313
---------------------------------------
14-
L1(avg): 0.402080 (0.0000)
15-
L2(avg): 0.272274 (0.0000)
16-
RMS(avg): 0.521799 (0.0000)
17-
Loss-fn(avg): 0.272274 (0.0000)
18-
R Squared: 0.652798 (0.0000)
14+
L1(avg): 3.472291 (0.0000)
15+
L2(avg): 26.064428 (0.0000)
16+
RMS(avg): 5.105333 (0.0000)
17+
Loss-fn(avg): 26.064428 (0.0000)
18+
R Squared: 0.998571 (0.0000)
1919

2020
---------------------------------------
2121
Physical memory usage(MB): %Number%
@@ -26,7 +26,7 @@ Virtual memory usage(MB): %Number%
2626
[1] 'Loading data for LightGBM' started.
2727
[1] 'Loading data for LightGBM' finished in %Time%.
2828
[2] 'Training with LightGBM' started.
29-
[2] (%Time%) Iteration: 50 Training-l2: 0.272273893168108
29+
[2] (%Time%) Iteration: 50 Training-l2: 26.0644295080124
3030
[2] 'Training with LightGBM' finished in %Time%.
3131
[3] 'Saving model' started.
3232
[3] 'Saving model' finished in %Time%.
Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,4 @@
1+
LightGBMR
2+
L1(avg) L2(avg) RMS(avg) Loss-fn(avg) R Squared /iter /lr /nl /mil /booster /v /nt Learner Name Train Dataset Test Dataset Results File Run Time Physical Memory Virtual Memory Command Line Settings
3+
3.472291 26.06443 5.105333 26.06443 0.998571 50 0.2 20 10 gbdt{l2=0.2 l1=0.2} + 1 LightGBMR %Data% %Data% %Output% 99 0 0 maml.exe TrainTest test=%Data% tr=LightGBMR{nt=1 iter=50 v=+ booster=gbdt{l1=0.2 l2=0.2} lr=0.2 mil=10 nl=20} dout=%Output% loader=Text{col=Label:R4:11 col=Features:R4:0-10 sep=; header+} data=%Data% out=%Output% seed=1 /iter:50;/lr:0.2;/nl:20;/mil:10;/booster:gbdt{l2=0.2 l1=0.2};/v:+;/nt:1
4+

0 commit comments

Comments
 (0)