Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 1 addition & 0 deletions llvm/docs/CommandGuide/llvm-exegesis.rst
Original file line number Diff line number Diff line change
Expand Up @@ -301,6 +301,7 @@ OPTIONS
* ``prepare-and-assemble-snippet``: Same as ``prepare-snippet``, but also dumps an excerpt of the sequence (hex encoded).
* ``assemble-measured-code``: Same as ``prepare-and-assemble-snippet``. but also creates the full sequence that can be dumped to a file using ``--dump-object-to-disk``.
* ``measure``: Same as ``assemble-measured-code``, but also runs the measurement.
* ``dry-run-measurement``: Same as measure, but does not actually execute the snippet.

.. option:: --x86-lbr-sample-period=<nBranches/sample>

Expand Down
11 changes: 11 additions & 0 deletions llvm/test/tools/llvm-exegesis/dry-run-measurement.test
Original file line number Diff line number Diff line change
@@ -0,0 +1,11 @@
# RUN: llvm-exegesis --mtriple=riscv64 --mcpu=sifive-p470 --mode=latency --opcode-name=ADD --use-dummy-perf-counters --benchmark-phase=dry-run-measurement | FileCheck %s
# REQUIRES: riscv-registered-target && native-registered-exegesis-target

# This test makes sure that llvm-exegesis doesn't execute "cross-compiled" snippets in the presence of
# --dry-run-measurement. RISC-V was chosen simply because most of the time we run tests on X86 machines.

# Should not contain misleading results.
# CHECK: measurements: []

# Should not contain error messages like "snippet crashed while running: Segmentation fault".
# CHECK: error: ''
6 changes: 6 additions & 0 deletions llvm/test/tools/llvm-exegesis/lit.local.cfg
Original file line number Diff line number Diff line change
Expand Up @@ -30,6 +30,12 @@ def can_use_perf_counters(mode, extra_options=[]):
print("could not exec llvm-exegesis")
return False

# LLJIT builds its own TargetMachine using arch designated by LLVM_TARGET_ARCH, which
# is default to host. We don't want tests that use LLJIT (but not necessarily
# execute the snippets) to run on machines that are not even supported by
# exegesis.
if config.root.native_target in ["AArch64", "Mips", "PowerPC", "RISCV", "X86"]:
config.available_features.add("native-registered-exegesis-target")

for arch in ["aarch64", "mips", "powerpc", "x86_64"]:
if can_execute_generated_snippets(arch):
Expand Down
1 change: 1 addition & 0 deletions llvm/tools/llvm-exegesis/lib/BenchmarkResult.h
Original file line number Diff line number Diff line change
Expand Up @@ -38,6 +38,7 @@ enum class BenchmarkPhaseSelectorE {
PrepareAndAssembleSnippet,
AssembleMeasuredCode,
Measure,
DryRunMeasure,
};

enum class BenchmarkFilter { All, RegOnly, WithMem };
Expand Down
33 changes: 25 additions & 8 deletions llvm/tools/llvm-exegesis/lib/BenchmarkRunner.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -99,22 +99,25 @@ class InProcessFunctionExecutorImpl : public BenchmarkRunner::FunctionExecutor {
static Expected<std::unique_ptr<InProcessFunctionExecutorImpl>>
create(const LLVMState &State, object::OwningBinary<object::ObjectFile> Obj,
BenchmarkRunner::ScratchSpace *Scratch,
std::optional<int> BenchmarkProcessCPU) {
std::optional<int> BenchmarkProcessCPU, bool DryRun) {
Expected<ExecutableFunction> EF =
ExecutableFunction::create(State.createTargetMachine(), std::move(Obj));

if (!EF)
return EF.takeError();

return std::unique_ptr<InProcessFunctionExecutorImpl>(
new InProcessFunctionExecutorImpl(State, std::move(*EF), Scratch));
new InProcessFunctionExecutorImpl(State, std::move(*EF), Scratch,
DryRun));
}

private:
InProcessFunctionExecutorImpl(const LLVMState &State,
ExecutableFunction Function,
BenchmarkRunner::ScratchSpace *Scratch)
: State(State), Function(std::move(Function)), Scratch(Scratch) {}
BenchmarkRunner::ScratchSpace *Scratch,
bool DryRun)
: State(State), Function(std::move(Function)), Scratch(Scratch),
DryRun(DryRun) {}

static void accumulateCounterValues(const SmallVector<int64_t, 4> &NewValues,
SmallVector<int64_t, 4> *Result) {
Expand Down Expand Up @@ -143,9 +146,14 @@ class InProcessFunctionExecutorImpl : public BenchmarkRunner::FunctionExecutor {
CrashRecoveryContext CRC;
CrashRecoveryContext::Enable();
const bool Crashed = !CRC.RunSafely([this, Counter, ScratchPtr]() {
Counter->start();
this->Function(ScratchPtr);
Counter->stop();
if (DryRun) {
Counter->start();
Counter->stop();
} else {
Counter->start();
this->Function(ScratchPtr);
Counter->stop();
}
});
CrashRecoveryContext::Disable();
PS.reset();
Expand Down Expand Up @@ -177,6 +185,7 @@ class InProcessFunctionExecutorImpl : public BenchmarkRunner::FunctionExecutor {
const LLVMState &State;
const ExecutableFunction Function;
BenchmarkRunner::ScratchSpace *const Scratch;
bool DryRun = false;
};

#ifdef __linux__
Expand Down Expand Up @@ -664,21 +673,29 @@ Expected<std::unique_ptr<BenchmarkRunner::FunctionExecutor>>
BenchmarkRunner::createFunctionExecutor(
object::OwningBinary<object::ObjectFile> ObjectFile,
const BenchmarkKey &Key, std::optional<int> BenchmarkProcessCPU) const {
bool DryRun =
BenchmarkPhaseSelector == BenchmarkPhaseSelectorE::DryRunMeasure;

switch (ExecutionMode) {
case ExecutionModeE::InProcess: {
if (BenchmarkProcessCPU.has_value())
return make_error<Failure>("The inprocess execution mode does not "
"support benchmark core pinning.");

auto InProcessExecutorOrErr = InProcessFunctionExecutorImpl::create(
State, std::move(ObjectFile), Scratch.get(), BenchmarkProcessCPU);
State, std::move(ObjectFile), Scratch.get(), BenchmarkProcessCPU,
DryRun);
if (!InProcessExecutorOrErr)
return InProcessExecutorOrErr.takeError();

return std::move(*InProcessExecutorOrErr);
}
case ExecutionModeE::SubProcess: {
#ifdef __linux__
if (DryRun)
return make_error<Failure>("The subprocess execution mode cannot "
"dry-run measurement at this moment.");

auto SubProcessExecutorOrErr = SubProcessFunctionExecutorImpl::create(
State, std::move(ObjectFile), Key, BenchmarkProcessCPU);
if (!SubProcessExecutorOrErr)
Expand Down
4 changes: 2 additions & 2 deletions llvm/tools/llvm-exegesis/lib/Target.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -98,7 +98,7 @@ ExegesisTarget::createBenchmarkRunner(
return nullptr;
case Benchmark::Latency:
case Benchmark::InverseThroughput:
if (BenchmarkPhaseSelector == BenchmarkPhaseSelectorE::Measure &&
if (BenchmarkPhaseSelector >= BenchmarkPhaseSelectorE::Measure &&
!PfmCounters.CycleCounter) {
const char *ModeName = Mode == Benchmark::Latency
? "latency"
Expand All @@ -116,7 +116,7 @@ ExegesisTarget::createBenchmarkRunner(
State, Mode, BenchmarkPhaseSelector, ResultAggMode, ExecutionMode,
ValidationCounters, BenchmarkRepeatCount);
case Benchmark::Uops:
if (BenchmarkPhaseSelector == BenchmarkPhaseSelectorE::Measure &&
if (BenchmarkPhaseSelector >= BenchmarkPhaseSelectorE::Measure &&
!PfmCounters.UopsCounter && !PfmCounters.IssueCounters)
return make_error<Failure>(
"can't run 'uops' mode, sched model does not define uops or issue "
Expand Down
9 changes: 6 additions & 3 deletions llvm/tools/llvm-exegesis/llvm-exegesis.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -132,7 +132,10 @@ static cl::opt<BenchmarkPhaseSelectorE> BenchmarkPhaseSelector(
clEnumValN(
BenchmarkPhaseSelectorE::Measure, "measure",
"Same as prepare-measured-code, but also runs the measurement "
"(default)")),
"(default)"),
clEnumValN(
BenchmarkPhaseSelectorE::DryRunMeasure, "dry-run-measurement",
"Same as measure, but does not actually execute the snippet")),
cl::init(BenchmarkPhaseSelectorE::Measure));

static cl::opt<bool>
Expand Down Expand Up @@ -476,7 +479,7 @@ static void runBenchmarkConfigurations(
}

void benchmarkMain() {
if (BenchmarkPhaseSelector == BenchmarkPhaseSelectorE::Measure &&
if (BenchmarkPhaseSelector >= BenchmarkPhaseSelectorE::Measure &&
!UseDummyPerfCounters) {
#ifndef HAVE_LIBPFM
ExitWithError(
Expand All @@ -501,7 +504,7 @@ void benchmarkMain() {

// Preliminary check to ensure features needed for requested
// benchmark mode are present on target CPU and/or OS.
if (BenchmarkPhaseSelector == BenchmarkPhaseSelectorE::Measure)
if (BenchmarkPhaseSelector >= BenchmarkPhaseSelectorE::Measure)
ExitOnErr(State.getExegesisTarget().checkFeatureSupport());

if (ExecutionMode == BenchmarkRunner::ExecutionModeE::SubProcess &&
Expand Down
Loading