55import ai .libs .jaicore .ml .core .dataset .serialization .ArffDatasetAdapter ;
66import ai .libs .jaicore .ml .core .filter .SplitterUtil ;
77import ai .libs .jaicore .ml .weka .classification .learner .IWekaClassifier ;
8- import ai .libs .mlplan .multiclass . wekamlplan .MLPlanWekaBuilder ;
8+ import ai .libs .mlplan .weka .MLPlanWekaBuilder ;
99import de .fraunhofer .iem .swan .model .MonteCarloValidator ;
1010import de .fraunhofer .iem .swan .util .Util ;
1111import org .api4 .java .ai .ml .core .dataset .schema .attribute .IAttribute ;
@@ -50,6 +50,7 @@ public MLPlan() {
5050 public HashMap <String , ArrayList <Double >> evaluateDataset (Instances instances1 ) {
5151
5252 String arffFilePath = Util .exportInstancesToArff (instances1 );
53+ ArffDatasetAdapter arffDatasetAdapter = new ArffDatasetAdapter ();
5354
5455 String mClass = Util .getClassName (instances1 );
5556
@@ -58,7 +59,7 @@ public HashMap<String, ArrayList<Double>> evaluateDataset(Instances instances1)
5859 //Initialize dataset using ARFF file path
5960 ILabeledDataset <?> dataset = null ;
6061 try {
61- dataset = ArffDatasetAdapter .readDataset (new File (arffFilePath ));
62+ dataset = arffDatasetAdapter .readDataset (new File (arffFilePath ));
6263 } catch (DatasetDeserializationFailedException e ) {
6364 e .printStackTrace ();
6465 }
@@ -95,14 +96,14 @@ public HashMap<String, ArrayList<Double>> evaluateDataset(Instances instances1)
9596 //optimizedClassifier.fit(split.get(0));
9697
9798 String trainPath = "swan/swan_core/swan-out/mlplan/train-methods-dataset.arff" ;
98- ArffDatasetAdapter .serializeDataset (new File (trainPath ), split .get (0 ));
99+ arffDatasetAdapter .serializeDataset (new File (trainPath ), split .get (0 ));
99100 ArffLoader trainLoader = new ArffLoader ();
100101 trainLoader .setFile (new File (trainPath ));
101102 Instances trainInstances = trainLoader .getDataSet ();
102103 trainInstances .setClassIndex (trainInstances .numAttributes () - 1 );
103104
104105 String testPath = "swan/swan_core/swan-out/mlplan/test-methods-dataset.arff" ;
105- ArffDatasetAdapter .serializeDataset (new File (testPath ), split .get (1 ));
106+ arffDatasetAdapter .serializeDataset (new File (testPath ), split .get (1 ));
106107 ArffLoader testLoader = new ArffLoader ();
107108 testLoader .setFile (new File (testPath ));
108109 Instances testInstances = testLoader .getDataSet ();
@@ -148,6 +149,7 @@ public void evaluateDataset(Instances instances, int k) {
148149
149150 //arffFilePath = "swan/swan_core/src/main/resources/waveform.arff";
150151 String arffFilePath = Util .exportInstancesToArff (instances );
152+ ArffDatasetAdapter arffDatasetAdapter = new ArffDatasetAdapter ();
151153
152154 String mClass = Util .getClassName (instances );
153155
@@ -156,7 +158,7 @@ public void evaluateDataset(Instances instances, int k) {
156158 //Initialize dataset using ARFF file path
157159 ILabeledDataset <?> dataset = null ;
158160 try {
159- dataset = ArffDatasetAdapter .readDataset (new File (arffFilePath ));
161+ dataset = arffDatasetAdapter .readDataset (new File (arffFilePath ));
160162 } catch (DatasetDeserializationFailedException e ) {
161163 e .printStackTrace ();
162164 }
@@ -177,7 +179,7 @@ public void evaluateDataset(Instances instances, int k) {
177179
178180 // System.out.println(attribute.getName());
179181 }
180- ArffDatasetAdapter .serializeDataset (new File ("swan/swan_core/swan-out/mlplan/methods-dataset.arff" ), split .get (1 ));
182+ arffDatasetAdapter .serializeDataset (new File ("swan/swan_core/swan-out/mlplan/methods-dataset.arff" ), split .get (1 ));
181183
182184
183185 for (int x = 0 ; x < split .get (1 ).size (); x ++) {
0 commit comments