Skip to content
Snippets Groups Projects
  1. Jun 11, 2022
    • Benjamin Maurer's avatar
      real/rsa run-time adjusted. · cdcd1370
      Benjamin Maurer authored
      Increased input size and added argument for iteration count.
      Run-times were: 0.095s/0.651s/3.257s
      Are now: 0,145s/1,551s/8,064s
      cdcd1370
  2. Jun 06, 2022
    • Benjamin Maurer's avatar
      Fixed stdout files for 'slow' mode. · 489b1b91
      Benjamin Maurer authored
      real/smallpt's .slowstdout file was missing.
      The value in spectral/dom-lt's was wrong, also runtime was lower
      than 5 sek on my machine (should be between 5-10s),
      so I increased the argument and updated dom-lt.slowstdout.
      489b1b91
  3. Jun 05, 2022
  4. Mar 10, 2022
  5. Mar 02, 2022
    • Alex D's avatar
      smallpt: truncate result · eabee5bd
      Alex D authored and Matthew Pickering's avatar Matthew Pickering committed
      Current test does some floating-point manipulations which can be non-deterministic
      due to different platforms/architectures. This change truncates the result to
      avoid potential stdout mismatch.
      eabee5bd
  6. Nov 13, 2021
  7. Oct 18, 2021
  8. May 07, 2021
  9. May 06, 2021
  10. Nov 19, 2020
  11. Nov 18, 2020
  12. Oct 13, 2020
    • Ben Gamari's avatar
      Introduce Shake-based build system · 612ad668
      Ben Gamari authored
      This introduces a Shake-based build system, building on Neil Mitchell's
      prototype from several years ago [1].
      
      This build system has a number of features relative to the existing
      make-based build system:
      
       * results are produced in a common CSV structure, making it easy to use
         the included analysis tool nofib-compare, to compare results from tests
         outside of nofib.
       * Cachegrind runs can be run in parallel, greatly speeding up such
         measurements
       * native support for (micro-)architectural performance counters on
         Linux via `perf stat`
      
      [1] https://gist.github.com/ndmitchell/4733855
      
      There are also a few more changes to tests to make this build system
      work well with nofib.
      
      In particular:
      
      * The Main module for some benchmarks has been renamed.
      * The boot phase of benchmarks is now done by a boot.sh script.
      * Replace some uses of perl with ghci.
      * Some benchmarks that we didn't use to run by default had
        missing/wrong(!) stdout files. I added/updated these.
      
      Take note that shake will run *more* benchmarks by default than make
      used to. So you might want to select a subset of benchmarks to run when
      using it.
      612ad668
  13. May 10, 2020
    • Simon Jakobi's avatar
      Fix build for real/eff/CSD · ee771193
      Simon Jakobi authored and Ben Gamari's avatar Ben Gamari committed
      Previously I would get an error during `make boot`:
      
          EffBench.hs:10:18: error:
              Could not load module ‘Control.Monad.State.Strict’
              It is a member of the hidden package ‘mtl-2.2.2’.
              You can run ‘:set -package mtl’ to expose it.
              (Note: this unloads all the modules in the current scope.)
              Use -v (or `:set -v` in ghci) to see a list of the files searched for.
             |
          10 | import qualified Control.Monad.State.Strict as S
             |
      ee771193
  14. Dec 20, 2019
  15. Nov 21, 2019
  16. Nov 14, 2019
  17. Mar 28, 2019
  18. Jan 15, 2019
  19. Jan 08, 2019
    • Sebastian Graf's avatar
      Stabilise benchmarks wrt. GC · 8632268a
      Sebastian Graf authored
      Summary:
      This is due to #15999, a follow-up on #5793 and #15357 and changes all
      benchmarks, some of them (i.e. `wheel-sieve1`, `awards`) rather drastically.
      
      The general plan is outlined in #15999: Identify GC-sensitive benchmarks by
      looking at how productivity rates change over different nursery sizes and
      iterate `main` of these benchmarks often enough for the non-monotony and
      discontinuities to go away.
      
      I was paying attention that the benchmarked logic is actually run $n times more
      often, rather than just benchmarking IO operations printing the result of CAFs.
      
      When I found benchmarks with insignificant runtime (#15357), I made sure that
      parameters/input files were adjusted so that runtime of the different modes
      fall within the ranges proposed in
      https://ghc.haskell.org/trac/ghc/ticket/15357#comment:4
      
      - fast: 0.1-0.2s
      - norm: 1-2s
      - slow: 5-10s
      
      This is what I did:
      
      - Stabilise bernoulli
      - Stabilise digits-of-e1
      - Stabilise digits-of-e2
      - Stabilise gen_regexp
      - Adjust running time of integrate
      - Adjust running time of kahan
      - Stabilise paraffins
      - Stabilise primes
      - Adjust running time of rfib
      - Adjust running time of tak
      - Stabilise wheel-sieve1
      - Stabilise wheel-sieve2
      - Adjust running time of x2n1
      - Adjust running time of ansi
      - Adjust running time of atom
      - Make awards benchmark something other than IO
      - Adjust running time of banner
      - Stabilise boyer
      - Adjust running time of boyer2
      - Adjust running time of queens
      - Adjust running time of calendar
      - Adjust runtime of cichelli
      - Stabilise circsim
      - Stabilise clausify
      - Stabilise constraints with moderate success
      - Adjust running time of cryptarithm1
      - Adjust running time of cryptarythm2
      - Adjust running time of cse
      - Adjust running time of eliza
      - Adjust running time of exact-reals
      - Adjust running time of expert
      - Stabilise fft2
      - Stabilise fibheaps
      - Stabilise fish
      - Adjust running time for gcd
      - Stabilise comp_lab_zift
      - Stabilise event
      - Stabilise fft
      - Stabilise genfft
      - Stabilise ida
      - Adjust running time for listcompr
      - Adjust running time for listcopy
      - Adjust running time of nucleic2
      - Attempt to stabilise parstof
      - Stabilise sched
      - Stabilise solid
      - Adjust running time of transform
      - Adjust running time of typecheck
      - Stabilise wang
      - Stabilise wave4main
      - Adjust running time of integer
      - Adjust running time of knights
      - Stabilise lambda
      - Stabilise lcss
      - Stabilise life
      - Stabilise mandel
      - Stabilise mandel2
      - Adjust running time of mate
      - Stabilise minimax
      - Adjust running time of multiplier
      - Adjust running time of para
      - Stabilise power
      - Adjust running time of primetest
      - Stabilise puzzle with mild success
      - Adjust running time for rewrite
      - Stabilise simple with mild success
      - Stabilise sorting
      - Stabilise sphere
      - Stabilise treejoin
      - Stabilise anna
      - Stabilise bspt
      - Stabilise cacheprof
      - Stablise compress
      - Stablise compress2
      - Stabilise fem
      - Adjust running time of fluid
      - Stabilise fulsom
      - Stabilise gamteb
      - Stabilise gg
      - Stabilise grep
      - Adjust running time of hidden
      - Stabilise hpg
      - Stabilise infer
      - Stabilise lift
      - Stabilise linear
      - Attempt to stabilise maillist
      - Stabilise mkhprog
      - Stabilise parser
      - Stabilise pic
      - Stabilise prolog
      - Attempt to stabilise reptile
      - Adjust running time of rsa
      - Adjust running time of scs
      - Stabilise symalg
      - Stabilise veritas
      - Stabilise binary-trees
      - Adjust running time of fasta
      - Adjust running time of k-nucleotide
      - Adjust running time of pidigits
      - Adjust running time of reverse-complement
      - Adjust running time of spectral-norm
      - Adjust running time of fannkuch-redux
      - Adjust running time for n-body
      
      Problematic benchmarks:
      
      - `last-piece`: Unclear how to stabilise. Runs for 300ms and I can't make up smaller inputs because I don't understand what it does.
      - `pretty`: It's just much too small to be relevant at all. Maybe we want to get rid of this one?
      - `scc`: Same as `pretty`. The input graph for which SCC analysis is done is much too small and I can't find good directed example graphs on the internet.
      - `secretary`: Apparently this needs `-package random` and consequently hasn't been run for a long time.
      - `simple`: Same as `last-piece`. Decent runtime (70ms), but it's unstable and I see no way to iterate it ~100 times in fast mode.
      - `eff`: Every benchmark is problematic here. Not from the point of view of allocations, but because the actual logic is vacuous. IMO, these should be performance tests, not actual benchmarks. Alternatively, write an actual application that makes use of algebraic effects.
      - `maillist`: Too trivial. It's just String/list manipulation, not representative of any Haskell code we would write today (no use of base library functions which could be fused, uses String instead of Text). It's only 75 loc according to `cloc`, that's not a `real` application.
      
      Reviewers: simonpj, simonmar, bgamari, AndreasK, osa1, alpmestan, O26 nofib
      
      GHC Trac Issues: #15999
      
      Differential Revision: https://phabricator.haskell.org/D5438
      8632268a
  20. Jan 01, 2019
  21. Dec 26, 2018
    • Sebastian Graf's avatar
      Compare results of compress by hashing · 042cf0be
      Sebastian Graf authored
      Summary:
      We don't want the result in the repo as it's a sizeable binary file that
      doesn't compress well.
      
      Storing the output file in the repository becomes infeasible for large
      inputs. There are two possible remedies:
      
      1. Generate the result files during `make boot` (Phab:D5426). We
         discovered some drawbacks (like missing dependency files to build
         `compress` during boot) to this approach which make it infeasible.
      2. Shrink the output files, for example by hashing the string that we would
         normally output and compare that instead.
      
      This patch implements the second alternative. This somewhat distorts the
      runtime profile, so we might want to consider doing hashing within the
      benchmark runner in the future.
      
      Test Plan: make boot
      
      Reviewers: AndreasK, nomeata, O26 nofib, osa1
      
      Reviewed By: AndreasK, osa1
      
      Subscribers: osa1
      
      Differential Revision: https://phabricator.haskell.org/D5469
      042cf0be
  22. Nov 22, 2018
    • Sebastian Graf's avatar
      Clean more generated files · f87d446b
      Sebastian Graf authored
      Summary:
      When we don't remove the generated files, `make boot` won't regenerate
      them. That in turn leads to expected output failures after bumping
      `{FAST,NORM,SLOW}_OPTS`.
      
      Test Plan: make clean && make boot && make
      
      Reviewers: AndreasK, mpickering, osa1, alpmestan, O26 nofib
      
      Reviewed By: alpmestan
      
      Differential Revision: https://phabricator.haskell.org/D5365
      f87d446b
  23. Nov 20, 2018
  24. Nov 14, 2018
    • Andreas Klebinger's avatar
      Adjust normal runtimes for nofib along with related changes · 08cc9b6b
      Andreas Klebinger authored
      Runtime for nofib benchmarks was all over the place.
      This patch adjusts runtime for most benchmarks such
      that it falls into the 0.2-2s range.
      
      This means that:
      * A default run will take longer
      * Time spent will be better distributed among benchmarks.
      * More benchmarks have runtimes long enough to be used
        for runtime analysis.
      
      Some more changes were done which go hand in hand
      with changing runtimes.
      * Some benchmarks now create their input files during boot.
      * Moved input files for anna in their own directory.
      * Remove printing of output for some of the floating
        point heavy benchmarks.
      * Added a comment about desired runtimes to README.
      * Made grep actually benchmark something.
      * Throw cachgrind out of the default benchmarks.
        The nondeterministic behaviour has been an issue for a
        while and it doesn't seem like an essential benchmark.
      
      Test Plan: run nofib in modes slow/normal/fast
      
      Reviewers: O26 nofib, alpmestan
      
      Reviewed By: alpmestan
      
      Subscribers: sgraf, alpmestan
      
      Differential Revision: https://phabricator.haskell.org/D4989
      08cc9b6b
Loading