Mercurial > dive4elements > river
diff artifacts/src/main/java/org/dive4elements/river/artifacts/sinfo/flowdepth/FlowDepthExporter.java @ 8863:1009cab0f86b
Some work on flow depth exporter
author | gernotbelger |
---|---|
date | Fri, 19 Jan 2018 18:47:53 +0100 |
parents | 7bbfb24e6eec |
children | 9f7a285b0ee3 |
line wrap: on
line diff
--- a/artifacts/src/main/java/org/dive4elements/river/artifacts/sinfo/flowdepth/FlowDepthExporter.java Fri Jan 19 18:47:18 2018 +0100 +++ b/artifacts/src/main/java/org/dive4elements/river/artifacts/sinfo/flowdepth/FlowDepthExporter.java Fri Jan 19 18:47:53 2018 +0100 @@ -10,7 +10,6 @@ import java.io.OutputStream; import java.text.DateFormat; -import java.text.NumberFormat; import java.util.ArrayList; import java.util.Collection; import java.util.Date; @@ -21,12 +20,13 @@ import org.apache.log4j.Logger; import org.dive4elements.artifacts.CallMeta; import org.dive4elements.artifacts.common.utils.Config; +import org.dive4elements.river.FLYS; import org.dive4elements.river.artifacts.model.CalculationResult; import org.dive4elements.river.artifacts.resources.Resources; import org.dive4elements.river.artifacts.sinfo.util.MetaAndTableJRDataSource; import org.dive4elements.river.exports.AbstractExporter; import org.dive4elements.river.model.River; -import org.dive4elements.river.utils.Formatter; +import org.dive4elements.river.model.Unit; import au.com.bytecode.opencsv.CSVWriter; import net.sf.jasperreports.engine.JRDataSource; @@ -56,61 +56,73 @@ private static final String CSV_LABEL_HEADER = "sinfo.export.flow_depth.csv.header.label"; private static final String CSV_GAUGE_HEADER = "sinfo.export.flow_depth.csv.header.gauge"; private static final String CSV_MEAN_BED_HEIGHT_HEADER = "sinfo.export.flow_depth.csv.header.mean_bed_height"; + private static final String CSV_MEAN_BED_HEIGHT_HEADER_SHORT = "sinfo.export.flow_depth.csv.header.mean_bed_height.short"; private static final String CSV_SOUNDING_HEADER = "sinfo.export.flow_depth.csv.header.sounding"; private static final String CSV_LOCATION_HEADER = "sinfo.export.flow_depth.csv.header.location"; private static final String CSV_META_HEADER_RESULT = "sinfo.export.flow_depth.csv.meta.header.result"; + private static final String CSV_META_HEADER_RESULT_LABEL = + "sinfo.export.flow_depth.csv.meta.header.result.label"; + private static final String CSV_META_VERSION = "sinfo.export.flow_depth.csv.meta.version"; + + private static final String CSV_META_VERSION_LABEL = + "sinfo.export.flow_depth.csv.meta.version.label"; private static final String CSV_META_USER = "sinfo.export.flow_depth.csv.meta.user"; + private static final String CSV_META_USER_LABEL = + "sinfo.export.flow_depth.csv.meta.user.label"; + private static final String CSV_META_CREATION = "sinfo.export.flow_depth.csv.meta.creation"; + private static final String CSV_META_CREATION_LABEL = + "sinfo.export.flow_depth.csv.meta.creation.label"; + private static final String CSV_META_RIVER = "sinfo.export.flow_depth.csv.meta.river"; + private static final String CSV_META_RIVER_LABEL = + "sinfo.export.flow_depth.csv.meta.river.label"; + private static final String CSV_META_HEADER_SOUNDING = "sinfo.export.flow_depth.csv.meta.header.sounding"; private static final String CSV_META_HEADER_WATERLEVEL = "sinfo.export.flow_depth.csv.meta.header.waterlevel"; + + private static final String CSV_META_RANGE = + "sinfo.export.flow_depth.csv.meta.range"; + + private static final String CSV_META_RANGE_LABEL = + "sinfo.export.flow_depth.csv.meta.range.label"; - private static final String JASPER_FILE = "/jasper/sinfo.flowdepth.jasper"; //$NON-NLS-1$ + private static final String CSV_META_HEIGHT_UNIT_RIVER = "sinfo.export.flow_depth.csv.meta.height_unit.river"; + + private static final String JASPER_FILE = "/jasper/sinfo.flowdepth.jasper"; + + private static final String UNIT_M = "m"; + + private static final String UNIT_CM = "cm"; + + private static final String UNIT_CUBIC_M = "m³/s"; /** The storage that contains the current calculation result.*/ private FlowDepthCalculationResults data = null; - private NumberFormat meanBedHeightFormatter; - - private NumberFormat tkhFormatter; - - private NumberFormat flowDepthFormatter; - - private NumberFormat getMeanBedHeightFormatter() { - if( meanBedHeightFormatter == null ) - // FIXME: check if this is right - meanBedHeightFormatter = Formatter.getMiddleBedHeightHeight(context); - return meanBedHeightFormatter; - } - - private NumberFormat getTkhFormatter() { - if( tkhFormatter == null ) - // FIXME: check if this is right, probably not, we need one digit - tkhFormatter = Formatter.getWaterlevelW(context); - return tkhFormatter; - } - - private NumberFormat getFlowDepthFormatter() { - if( flowDepthFormatter == null ) - // FIXME: check if this is right - flowDepthFormatter = Formatter.getMeterFormat(context); - return flowDepthFormatter; - } + /** + * Formats header with unit + */ + private String msgUnit(final String key, String unit ) { + + final String msg = msg(key); + return String.format("%s [%s]", msg, unit); + } @Override protected void addData(Object d) { @@ -124,7 +136,7 @@ data = (FlowDepthCalculationResults)dat; } } - + @Override protected void writeCSVData(CSVWriter writer) { log.info("FlowDepthExporter.writeCSVData"); @@ -132,17 +144,12 @@ /* fetch calculation results */ final FlowDepthCalculationResults results = data; - /* write as csv */ + final boolean useTkh = results.isUseTkh(); + final River river = results.getRiver(); -// boolean atGauge = mode == WQ_MODE.QGAUGE || mode == WQ_MODE.WGAUGE; -// boolean isQ = mode == WQ_MODE.QGAUGE || mode == WQ_MODE.QFREE; -// RiverUtils.WQ_INPUT input -// = RiverUtils.getWQInputMode((D4EArtifact)master); - - final boolean useTkh = results.isUseTkh(); - + /* write as csv */ writeCSVMeta(writer, results); - writeCSVHeader(writer, useTkh); + writeCSVHeader(writer, river, useTkh); for (final FlowDepthCalculationResult result : results.getResults()) { writeCSVFlowDepthResult(writer, result, useTkh); @@ -159,49 +166,35 @@ private void writeCSVMeta(final CSVWriter writer, final FlowDepthCalculationResults results) { log.info("FlowDepthExporter.writeCSVMeta"); - // Workflow zur Berechnung der Fließtiefe.pdf - // "##ERGEBNISAUSGABE - Name des Gewässers - Fließtiefe" + final String calcModeLabel = results.getCalcModeLabel(); final River river = results.getRiver(); - writeCSVMeataEntry(writer, CSV_META_HEADER_RESULT, river.getName() ); + writeCSVMetaEntry(writer, CSV_META_HEADER_RESULT, msg( CSV_META_HEADER_RESULT_LABEL ), river.getName(), calcModeLabel ); // "# FLYS-Version: " - // FIXME - final String flysVersion = "unbekannt"; - writeCSVMeataEntry(writer, CSV_META_VERSION, flysVersion ); + writeCSVMetaEntry(writer, CSV_META_VERSION, msg( CSV_META_VERSION_LABEL ), FLYS.VERSION ); // "# Bearbeiter: " - // FIXME - final String user = "unbekannt"; - writeCSVMeataEntry(writer, CSV_META_USER, user ); + writeCSVMetaEntry(writer, CSV_META_USER, msg( CSV_META_USER_LABEL ), results.getUser() ); // "# Datum der Erstellung: " final Locale locale = Resources.getLocale(context.getMeta()); final DateFormat df = DateFormat.getDateInstance(DateFormat.SHORT, locale); - writeCSVMeataEntry(writer, CSV_META_CREATION, df.format(new Date()) ); + writeCSVMetaEntry(writer, CSV_META_CREATION, msg( CSV_META_CREATION_LABEL ), df.format(new Date()) ); // "# Gewässer: " - writeCSVMeataEntry(writer, CSV_META_RIVER, river.getName() ); + writeCSVMetaEntry(writer, CSV_META_RIVER, msg( CSV_META_RIVER_LABEL ), river.getName() ); // "# Höhensystem des Flusses: " - - // FIXME + final Unit wstUnit = river.getWstUnit(); + writeCSVMetaEntry(writer, CSV_META_HEIGHT_UNIT_RIVER, wstUnit.getName()); // "# Ort/Bereich (km): " - // FIXME - // TODO: unklar, es wird nur ein Bereich eingegeben -// RangeAccess rangeAccess = new RangeAccess(flys); -// double[] kms = rangeAccess.getKmRange(); -// writer.writeNext(new String[] { -// Resources.getMsg( -// meta, -// CSV_META_RANGE, -// CSV_META_RANGE, -// new Object[] { kms[0], kms[kms.length-1] }) -// }); + writeCSVMetaEntry(writer, CSV_META_RANGE, msg( CSV_META_RANGE_LABEL ), getKmFormatter().format(results.getFrom() ), getKmFormatter().format( results.getTo())); // "##METADATEN PEILUNG" - writeCSVMeataEntry(writer, CSV_META_HEADER_SOUNDING ); + writeCSVMetaEntry(writer, CSV_META_HEADER_SOUNDING ); + // FIXME: check: macht nicht viel sinn da es mehrere geben kann.. oder immer wieder wiederholen? // "# Jahr der Peilung: " // FIXME // "# Aufnahmeart: " @@ -213,7 +206,7 @@ // "# ursprüngliches Höhensystem: " // FIXME // "##METADATEN WASSERSPIEGELLAGE" - writeCSVMeataEntry(writer, CSV_META_HEADER_WATERLEVEL ); + writeCSVMetaEntry(writer, CSV_META_HEADER_WATERLEVEL ); // "# Bezeichnung der Wasserspiegellage: " // FIXME // "# Höhensystem der Wasserspiegellage: " @@ -231,19 +224,11 @@ // "# Q (m³/s): " (nur bei Eingabe des Durchflusses) // TODO: unklar, es wird kein Q eingegeben -// writer.writeNext(new String[] { -// Resources.getMsg( -// meta, -// CSV_META_GAUGE, -// CSV_META_GAUGE, -// new Object[] { RiverUtils.getGaugename(flys) }) -// }); - writer.writeNext(new String[] { "" }); } - private void writeCSVMeataEntry(CSVWriter writer, String message, Object... messageArgs) { + private void writeCSVMetaEntry(CSVWriter writer, String message, Object... messageArgs) { CallMeta meta = context.getMeta(); @@ -259,28 +244,32 @@ /** * Write the header, with different headings depending on whether at a * gauge or at a location. + * @param river * @param useTkh */ private void writeCSVHeader( final CSVWriter writer, + final River river, final boolean useTkh ) { log.info("FlowDepthExporter.writeCSVHeader"); final Collection<String> header = new ArrayList<>(11); - header.add(msg(CSV_KM_HEADER,CSV_KM_HEADER)); - header.add(msg(CSV_FLOWDEPTH_HEADER)); + header.add(msg(CSV_KM_HEADER)); + header.add(msgUnit(CSV_FLOWDEPTH_HEADER, UNIT_M)); if( useTkh ) { - header.add(msg(CSV_FLOWDEPTHTKH_HEADER)); - header.add(msg(CSV_TKH_HEADER)); + header.add(msgUnit(CSV_FLOWDEPTHTKH_HEADER, UNIT_M)); + header.add(msgUnit(CSV_TKH_HEADER, UNIT_CM)); } - header.add(msg(CSV_WATERLEVEL_HEADER)); - header.add(msg(CSV_DISCHARGE_HEADER)); + + final String wstUnitName = river.getWstUnit().getName(); + header.add(msgUnit(CSV_WATERLEVEL_HEADER, wstUnitName)); + header.add(msgUnit(CSV_DISCHARGE_HEADER, UNIT_CUBIC_M)); header.add(msg(CSV_LABEL_HEADER)); header.add(msg(CSV_GAUGE_HEADER)); - header.add(msg(CSV_MEAN_BED_HEIGHT_HEADER)); + header.add(msgUnit(CSV_MEAN_BED_HEIGHT_HEADER, wstUnitName)); header.add(msg(CSV_SOUNDING_HEADER)); header.add(msg(CSV_LOCATION_HEADER)); @@ -313,7 +302,7 @@ } // Wasserstand [NN + m] - lines.add( getWFormatter().format( row.getWaterlevel() ) ); + lines.add( getW2Formatter().format( row.getWaterlevel() ) ); // Q [m³/s] lines.add( getQFormatter().format( row.getDischarge() ) ); @@ -325,7 +314,7 @@ lines.add( row.getGauge() ); // Mittlere Sohlhöhe [NN + m] - lines.add( getMeanBedHeightFormatter().format( row.getMeanBedHeight( ) ) ); + lines.add( getMeanBedHeighFormatter().format( row.getMeanBedHeight( ) ) ); // Peilung/Epoche lines.add( row.getSoundageLabel() ); @@ -394,52 +383,50 @@ private void addJRMetaData(final MetaAndTableJRDataSource source, FlowDepthCalculationResults results) { - // Workflow zur Berechnung der Fließtiefe.pdf - // "##ERGEBNISAUSGABE - Name des Gewässers - Fließtiefe" - // writeCSVMeataEntry(writer, CSV_META_HEADER_RESULT, inputData.getRiver() ); - - // FIXME - final String flysVersion = "unbekannt"; - // CSV_META_VERSION - source.addMetaData("version", flysVersion); + final River river = results.getRiver(); + final String wstUnitName = river.getWstUnit().getName(); - // FIXME - String user = "unbekannt"; - // CSV_META_USER - source.addMetaData("user", user); + /* general metadata */ + source.addMetaData("header", msg(CSV_META_HEADER_RESULT_LABEL)); + source.addMetaData("calcMode", results.getCalcModeLabel()); + + source.addMetaData("version_label", msg(CSV_META_VERSION_LABEL)); + source.addMetaData("version", FLYS.VERSION); + + source.addMetaData("user_label", msg(CSV_META_USER_LABEL)); + source.addMetaData("user", results.getUser()); final Locale locale = Resources.getLocale(context.getMeta()); final DateFormat df = DateFormat.getDateInstance(DateFormat.SHORT, locale); + source.addMetaData("date_label", msg(CSV_META_CREATION_LABEL)); source.addMetaData("date", df.format(new Date())); - // CSV_META_RIVER - source.addMetaData("river", results.getRiver().getName()); + source.addMetaData("river_label", msg(CSV_META_RIVER_LABEL) ); + source.addMetaData("river", river.getName()); - // FIXME - source.addMetaData("range", "FIXME"); - // "# Ort/Bereich (km): " - // FIXME - // TODO: unklar, es wird nur ein Bereich eingegeben -// RangeAccess rangeAccess = new RangeAccess(flys); -// double[] kms = rangeAccess.getKmRange(); -// writer.writeNext(new String[] { -// Resources.getMsg( -// meta, -// CSV_META_RANGE, -// CSV_META_RANGE, -// new Object[] { kms[0], kms[kms.length-1] }) -// }); + final String rangeValue = String.format( "%s - %s", getKmFormatter().format(results.getFrom() ), getKmFormatter().format( results.getTo())); + source.addMetaData("range_label", msg(CSV_META_RANGE_LABEL)); + source.addMetaData("range", rangeValue); -// RangeAccess rangeAccess = new RangeAccess(flys); -// double[] kms = rangeAccess.getKmRange(); -// source.addMetaData("range", -// kmf.format(kms[0]) + " - " + kmf.format(kms[kms.length-1])); + /* column headings */ + source.addMetaData("station_header", msg(CSV_KM_HEADER)); + source.addMetaData("flowdepth_header", msg(CSV_FLOWDEPTH_HEADER)); + source.addMetaData("flowdepth_tkh_header", msg(CSV_FLOWDEPTHTKH_HEADER)); + source.addMetaData("tkh_header", msg(CSV_TKH_HEADER)); + source.addMetaData("waterlevel_header", msg(CSV_WATERLEVEL_HEADER)); + source.addMetaData("river_unit", wstUnitName); + source.addMetaData("discharge_header", msg(CSV_DISCHARGE_HEADER)); + source.addMetaData("waterlevel_name_header", msg(CSV_LABEL_HEADER)); + source.addMetaData("gauge_header", msg(CSV_GAUGE_HEADER)); + source.addMetaData("bedheight_header", msg(CSV_MEAN_BED_HEIGHT_HEADER_SHORT)); + source.addMetaData("sounding_name_header", msg(CSV_SOUNDING_HEADER)); + source.addMetaData("location_header", msg(CSV_LOCATION_HEADER)); } private void addJRTableData(final MetaAndTableJRDataSource source, final FlowDepthCalculationResult result, final boolean useTkh) { final Collection<FlowDepthRow> rows = result.getRows(); - + for (final FlowDepthRow row : rows) { final String[] formattedRow = formatFlowDepthRow(row, useTkh);