Mercurial > dive4elements > river
diff artifacts/src/main/java/org/dive4elements/river/artifacts/sinfo/flowdepth/FlowDepthExporter.java @ 9150:23945061daec
gigantic refactoring: exporter, result, results
to support multiple jaspers -> collisions
author | gernotbelger |
---|---|
date | Thu, 14 Jun 2018 16:56:31 +0200 |
parents | 7134a4c7d1b6 |
children | a4121ec450d6 |
line wrap: on
line diff
--- a/artifacts/src/main/java/org/dive4elements/river/artifacts/sinfo/flowdepth/FlowDepthExporter.java Thu Jun 14 15:12:25 2018 +0200 +++ b/artifacts/src/main/java/org/dive4elements/river/artifacts/sinfo/flowdepth/FlowDepthExporter.java Thu Jun 14 16:56:31 2018 +0200 @@ -9,20 +9,14 @@ package org.dive4elements.river.artifacts.sinfo.flowdepth; import java.io.OutputStream; -import java.util.ArrayList; -import java.util.Collection; -import org.apache.log4j.Logger; -import org.dive4elements.river.artifacts.common.GeneralResultType; +import org.dive4elements.river.artifacts.common.AbstractCalculationExportableResult; +import org.dive4elements.river.artifacts.common.AbstractCommonExporter; +import org.dive4elements.river.artifacts.common.ExportContextCSV; import org.dive4elements.river.artifacts.common.JasperDesigner; import org.dive4elements.river.artifacts.common.JasperReporter; import org.dive4elements.river.artifacts.common.MetaAndTableJRDataSource; -import org.dive4elements.river.artifacts.common.ResultRow; -import org.dive4elements.river.artifacts.sinfo.common.AbstractSInfoExporter; -import org.dive4elements.river.artifacts.sinfo.common.SInfoResultType; -import org.dive4elements.river.artifacts.sinfo.util.BedHeightInfo; import org.dive4elements.river.artifacts.sinfo.util.RiverInfo; -import org.dive4elements.river.artifacts.sinfo.util.WstInfo; import au.com.bytecode.opencsv.CSVWriter; import net.sf.jasperreports.engine.JRException; @@ -34,139 +28,67 @@ * @author Gernot Belger */ // REMARK: must be public because its registered in generators.xml -public class FlowDepthExporter extends AbstractSInfoExporter<FlowDepthCalculationResult, FlowDepthCalculationResults> { - - /** The log used in this exporter. */ - private static Logger log = Logger.getLogger(FlowDepthExporter.class); - - private static final String JASPER_FILE = "/jasper/templates/sinfo.flowdepth.jrxml"; // TODO use jrxml-path all over the project +public class FlowDepthExporter extends AbstractCommonExporter<FlowDepthCalculationResults> { @Override - protected Logger getLog() { - return log; - } + protected void doWriteCSVData(final CSVWriter writer, final FlowDepthCalculationResults results) { + // TODO: Diesen Ablauf in super? - @Override - protected void writeCSVGlobalMetadata(final CSVWriter writer, final FlowDepthCalculationResults results) { - log.info("FlowDepthExporter.writeCSVMeta"); - super.writeCSVGlobalMetadataDefaults(writer, results); + // TODO: move results into context? + final ExportContextCSV exportContextCSV = new ExportContextCSV(this.context, writer); + + getLog().info("writeCSVData"); + + /* write as csv */ + exportContextCSV.writeCSVGlobalMetadataDefaults(results); // ggf auslagern innerhalb dieser Klasse + + // writer.writeNext(new String[] { "" }); // break line HERE to avoid redundance + + final RiverInfo river = results.getRiver(); + + final Class<?> lastResultType = null; + + for (final AbstractCalculationExportableResult<FlowDepthCalculationResults> result : results.getResults()) { + + final Class<?> resultType = result.getClass(); + if (lastResultType == null || lastResultType != resultType) { + exportContextCSV.writeBlankLine(); + result.writeCSVHeader(exportContextCSV, results, river); + exportContextCSV.writeBlankLine(); + } else + exportContextCSV.writeCSVLine(new String[] { "#" }); + + result.writeCsv(exportContextCSV, results); + } } @Override - protected void writeCSVResultMetadata(final CSVWriter writer, final FlowDepthCalculationResults results, final FlowDepthCalculationResult result) { - - final BedHeightInfo sounding = result.getSounding(); - super.writeCSVSoundingMetadata(writer, sounding); - - writer.writeNext(new String[] { "" }); // break line - - final WstInfo wst = result.getWst(); - super.writeCSVWaterlevelMetadata(writer, wst); - - writer.writeNext(new String[] { "" }); // break line - - } - - /** - * Write the header, with different headings depending on whether at a - * gauge or at a location. - * - * @param river - * @param useTkh - */ - @Override - protected void writeCSVHeader(final CSVWriter writer, final FlowDepthCalculationResults results, final RiverInfo river) { - log.info("FlowDepthExporter.writeCSVHeader"); - - final Collection<String> header = new ArrayList<>(11); - - header.add(msg(GeneralResultType.station.getCsvHeader())); - header.add(msgUnit(SInfoResultType.flowdepth.getCsvHeader(), SInfoResultType.flowdepth.getUnit())); - - if (results.isUseTkh()) { - header.add(msgUnit(SInfoResultType.flowdepthtkh.getCsvHeader(), SInfoResultType.flowdepthtkh.getUnit())); - header.add(msgUnit(SInfoResultType.tkh.getCsvHeader(), SInfoResultType.tkh.getUnit())); - } - - header.add(msgUnit(SInfoResultType.waterlevel.getCsvHeader(), river.getWstUnit())); - header.add(msgUnit(SInfoResultType.discharge.getCsvHeader(), SInfoResultType.discharge.getUnit())); - header.add(msg(SInfoResultType.waterlevelLabel.getCsvHeader())); - header.add(msg(SInfoResultType.gaugeLabel.getCsvHeader())); - header.add(msgUnit(SInfoResultType.meanBedHeight.getCsvHeader(), river.getWstUnit())); - header.add(msg(SInfoResultType.soundingLabel.getCsvHeader())); - header.add(msg(SInfoResultType.location.getCsvHeader())); + protected void doWritePdf(final OutputStream out, final FlowDepthCalculationResults results) { - writer.writeNext(header.toArray(new String[header.size()])); - } - - /** - * Format a row of a flow depth result into an array of string, both used by csv and pdf - * - * @param pdf - * - * @param useTkh - */ - @Override - protected String[] formatRow(final FlowDepthCalculationResults results, final ResultRow row, final ExportMode mode) { - - final Collection<String> lines = new ArrayList<>(11); - - lines.add(row.exportValue(this.context, GeneralResultType.station)); - lines.add(row.exportValue(this.context, SInfoResultType.flowdepth)); - - if (mode == ExportMode.pdf || getData().isUseTkh()) { - lines.add(row.exportValue(this.context, SInfoResultType.flowdepthtkh)); - lines.add(row.exportValue(this.context, SInfoResultType.tkh)); - } + // TODO: Move to super? Maybe not, hier gibt es unterschiede -> design remove columns + try { + final ExportContextCSV exportContextCSV = new ExportContextCSV(this.context, null); - lines.add(row.exportValue(this.context, SInfoResultType.waterlevel)); - lines.add(row.exportValue(this.context, SInfoResultType.discharge)); - lines.add(row.exportValue(this.context, SInfoResultType.waterlevelLabel)); - lines.add(row.exportValue(this.context, SInfoResultType.gaugeLabel)); - lines.add(row.exportValue(this.context, SInfoResultType.meanBedHeight)); - lines.add(row.exportValue(this.context, SInfoResultType.soundingLabel)); - lines.add(row.exportValue(this.context, SInfoResultType.location)); - - return lines.toArray(new String[lines.size()]); - } + final JasperReporter reporter = new JasperReporter(); - @Override - protected void writePDF(final OutputStream out) { + for (final AbstractCalculationExportableResult<FlowDepthCalculationResults> result : results.getResults()) { + final MetaAndTableJRDataSource source = new MetaAndTableJRDataSource(); + getHelper().addJRMetaDataUSINFO(source, results); - try { - final MetaAndTableJRDataSource source = createJRData(this.data); - final JasperReporter reporter = new JasperReporter(); - final JasperDesigner design = reporter.addReport(JASPER_FILE, source); - - if (!getData().isUseTkh()) { - design.removeColumn("tkh"); - design.removeColumn("flowdepthtkh"); + final JasperDesigner design = result.addReport(exportContextCSV, results, reporter, source); + if (result instanceof FlowDepthCalculationResult) { + if (!((FlowDepthCalculationResult) result).hasTkh()) { + design.removeColumn("tkh"); + design.removeColumn("flowdepthtkh"); + } + } + reporter.exportPDF(out); } - reporter.exportPDF(out); } catch (final JRException je) { getLog().warn("Error generating PDF Report!", je); } - } - - @Override - protected final void addJRMetaData(final MetaAndTableJRDataSource source, final FlowDepthCalculationResults results) { - - /* general metadata */ - super.addJRMetaData(source, results); - /* column headings */ - source.addMetaData("station_header", GeneralResultType.station.getPdfHeader(this.context.getMeta())); - source.addMetaData("flowdepth_header", SInfoResultType.flowdepth.getPdfHeader(this.context.getMeta())); - source.addMetaData("flowdepth_tkh_header", SInfoResultType.flowdepthtkh.getPdfHeader(this.context.getMeta())); - source.addMetaData("tkh_header", SInfoResultType.tkh.getPdfHeader(this.context.getMeta())); - source.addMetaData("waterlevel_header", SInfoResultType.waterlevel.getPdfHeader(this.context.getMeta())); - source.addMetaData("discharge_header", SInfoResultType.discharge.getPdfHeader(this.context.getMeta())); - source.addMetaData("waterlevel_name_header", SInfoResultType.waterlevelLabel.getPdfHeader(this.context.getMeta())); - source.addMetaData("gauge_header", SInfoResultType.gaugeLabel.getPdfHeader(this.context.getMeta())); - source.addMetaData("bedheight_header", SInfoResultType.meanBedHeight.getPdfHeader(this.context.getMeta())); - source.addMetaData("sounding_name_header", SInfoResultType.soundingLabel.getPdfHeader(this.context.getMeta())); - source.addMetaData("location_header", SInfoResultType.location.getPdfHeader(this.context.getMeta())); } } \ No newline at end of file