Re: Continuous benchmarking setup
Thanks Tom and Antoine!
Since these benchmarks are literally running on a machine in my closet
at home, there may be some downtime in the future. At some point we
should document a process of setting up a new machine from scratch to
be the nightly bare metal benchmark slave.
On Fri, May 11, 2018 at 9:08 AM, Antoine Pitrou <solipsis@xxxxxxxxxx> wrote:
> Hi again,
> Tom has configured the benchmarking machine to run and publish Arrow's
> ASV-based benchmarks. The latest results can now be seen at:
> I expect these are regenerated on a regular (daily?) basis.
> Thanks Tom :-)
> On Wed, 11 Apr 2018 15:40:17 +0200
> Antoine Pitrou <antoine@xxxxxxxxxx> wrote:
>> With the following changes, it seems we might reach the point where
>> we're able to run the Python-based benchmark suite accross multiple
>> commits (at least the ones not anterior to those changes):
>> To make this truly useful, we would need a dedicated host. Ideally a
>> (Linux) OS running on bare metal, with SMT/HyperThreading disabled.
>> If running virtualized, the VM should have dedicated physical CPU cores.
>> That machine would run the benchmarks on a regular basis (perhaps once
>> per night) and publish the results in static HTML form somewhere.
>> (note: nice to have in the future might be access to NVidia hardware,
>> but right now there are no CUDA benchmarks in the Python benchmarks)
>> What should be the procedure here?