Publishing details
Changelog
python-oq-engine (2.4.0-0~trusty01) trusty; urgency=low
[Michele Simionato]
* Now the command `oq export loss_curves/rlz-XXX` works both for the
`classical_risk` calculator and the `event_based_risk` calculator
[Daniele Viganò]
* Remove the default 30 day-old view limit in the WebUI calculation list
[Michele Simionato]
* Fixed a broken import affecting the command `oq upgrade_nrml`
* Made it possible to specify multiple file names in <uncertaintyValue/>
in the source_model_logic_tree file
* Reduced the data transfer in the object `RlzsAssoc` and improved the
postprocessing of hazard curves when the option `--hc` is given
* Changed the `ruptures.xml` exporter to export unique ruptures
* Fixed a bug when downloading the outputs from the WebUI on Windows
* Made `oq info --report` fast again by removing the rupture fine filtering
* Improved the readibility of the CSV export `dmg_total`
* Removed the column `eid` from the CSV export `ruptures`; also
renamed the field `serial` to `rup_id` and reordered the fields
* Changed the event loss table exporter: now it exports an additional
column with the `rup_id`
* Changed scenario npz export to export also the GMFs outside the maximum
distance
* Fixed scenario npz export when there is a single event
* Replaced the event tags with numeric event IDs
* The mean hazard curves are now generated by default
* Improved the help message of the command `oq purge`
* Added a `@reader` decorator to mark tasks reading directly from the
file system
* Removed the .txt exporter for the GMFs, used internally in the tests
* Fixed a bug with relative costs which affected master for a long time,
but not the release 2.3. The insured losses were wrong in that case.
* Added an .hdf5 exporter for the asset loss table
* Loss maps and aggregate losses are computed in parallel or sequentially
depending if the calculation is a postprocessing calculation or not
* Deprecated the XML risk exporters
* Removed the .ext5 file
* Restored the parameter `asset_loss_table` in the event based calculators
* Added a full .hdf5 exporter for `hcurves-rlzs`
* Removed the `individual_curves` flag: now by default only the statistical
hazard outputs are exported
* Saved *a lot* of memory in the computation of the hazard curves and stats
* Renamed the parameter `all_losses` to `asset_loss_table`
* Added an experimental version of the event based risk calculator which
is able to use GMFs imported from an external file
* Added a `max_curve` functionality to compute the upper limit of the
hazard curves amongst realizations
* Raised an error if the user specifies `quantile_loss_curves`
or `conditional_loss_poes` in a classical_damage calculation
* Added a CSV exporter for the benefit-cost-ratio calculator
* The classical_risk calculator now reads directly the probability maps,
not the hazard curves
* Turned the loss curves into on-demand outputs
for the event based risk calculator
* The loss ratios are now stored in the datastore and not in an
external .ext5 file
* The engine outputs are now streamed by the WebUI
* Used a temporary export directory in the tests, to avoid conflicts
in multiuser situations
* Added an .npz exporter for the loss maps
* Raised an error early when using a complex logic tree in scenario
calculations
* Changed the CSV exporter for the loss curves: now it exports all the
curves for a given site for the classical_risk calculator
* Fixed the save_ruptures procedure when there are more than 256
surfaces in the MultiSurface
* Renamed the `csq_` outputs of the scenario_damage to `losses_`
* Changed the way scenario_damage are stored internally to be more
consistent with the other calculators
* Removed the GSIM from the exported file name of the risk outputs
* New CSV exporter for GMFs generated by the event based calculator
* The event IDs are now unique and a constraint on the maximum
number of source groups (65,536) has been added
* Added an output `losses_by_event` to the scenario_risk calculator
* Changed the output `ruptures.csv` to avoid duplications
* Added an output `losses_by_taxon` to the scenario_risk calculator
* Fixed a performance bug in `get_gmfs`: now the scenario risk and damage
calculators are orders of magnitude faster for big arrays
* Added an export test for the event loss table in the case of multiple TRTs
* Removed the experimental `rup_data` output
* Added an .npz export for the output `losses_by_asset`
* Exported the scenario_risk aggregate losses in a nicer format
[Daniele Viganò]
* The 'oq webui' command now works on a multi-user installation
* Splitted RPM packages into python-oq-engine (single node)and
python-oq-engine-master/python-oq-engine-worker (multi-node)
[Paolo Tormene]
* The 'Continue' button in the Web UI is now available also for risk
calculations
[Michele Simionato]
* Fixed a Python 3 bug in the WebUI when continuing a calculation: the
hazard_calculation_id was passed as a string and not as an integer
* Changed to rupture storage to use variable length-arrays, with a speedup
of two orders of magnitude
* Avoided storing twice the rupture events
* Optimized the serialization of ruptures on HDF5 by using a `sids` output
* Changed the Web UI button from "Run Risk" to "Continue"
* The `avg` field in the loss curves is computed as the integral of the curve
again, and it is not extracted from the avg_losses output anymore
* Made the `fullreport` exportable
* Fixed the `rup_data` export, since the boundary field was broken
* Restored the output `losses_by_taxon` in the event_based_risk calculator
* Fixed the calculator event based UCERF so that average losses can
be stored
[Daniele Viganò]
* Added a check to verify that an 'oq' client is talking to the
right DbServer instance
* Introduced an optional argument for 'oq dbserver' command line
to be able to override its default interface binding behaviour
[Michele Simionato]
* Optimized the event based calculators by reducing the number of calls
to the GmfComputer and by using larger arrays
* Added a check on missing vulnerability functions for some loss type
for some taxonomy
* Now we save the GMFs on the .ext5 file, not the datastore
* Fixed bug in event_based_risk: it was impossible to use vulnerability
functions with "PM" distribution
* Fixed bug in event_based_risk: the ebrisk calculator is required as
precalculator of event_based_risk, not others
* Fixed bug in scenario_risk: the output `all_losses-rlzs` was aggregated
incorrectly
* Now the ucerf_risk calculators transfer only the events, not the ruptures,
thus reducing the data transfer of several orders of magnitude
* Added a view `get_available_gsims` to the WebUI and fixed the API docs
* Introduced a configuration parameter `max_site_model_distance` with default
of 5 km
* Implemented sampling in the UCERF event based hazard calculator
[Daniele Viganò]
* Use threads instead of processes in DbServer because SQLite3
isn't fork-safe on macOS Sierra
[Michele Simionato]
* Fixed a TypeError when deleting a calculation from the WebUI
* Extended the command `oq to_hdf5` to manage source model files too
* Improved significantly the performance of the event based calculator
when computing the GMFs and not the hazard curves
* Stored information about the mean ground motion in the datastore
* Saved the rupture mesh with 32 floats instead of 64 bit floats
* Raised the limit on the event IDs from 2^16 to 2^32 per task
* Fixed classical_risk: there was an error when computing the statistics
in the case of multiple assets of the same taxonomy on the same site
* Changed the UCERF event based calculators to parallelize by SES
* Fixed a site model bug: when the sites are extracted from the site model
there is no need to perform geospatial queries to get the parameters
* Added a command `oq normalize` to produce good `sites.csv` files
* Introduced a `ses_seed` parameter to specify the seed used to generate
the stochastic event sets; `random_seed` is used for the sampling only
* Changed the `build_rcurves` procedure to read the loss ratios directly from
the workers
-- Matteo Nastasi (GEM Foundation) <email address hidden> Tue, 23 May 2017 10:46:56 +0200
Builds
Built packages
-
python-oq-engine
computes seismic hazard and physical risk
Package files