You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
742 lines
29 KiB
742 lines
29 KiB
/*
|
|
* Copyright (C) 2014 The Android Open Source Project
|
|
*
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
#ifndef ART_COMPILER_OPTIMIZING_OPTIMIZING_UNIT_TEST_H_
|
|
#define ART_COMPILER_OPTIMIZING_OPTIMIZING_UNIT_TEST_H_
|
|
|
|
#include <memory>
|
|
#include <ostream>
|
|
#include <string_view>
|
|
#include <string>
|
|
#include <tuple>
|
|
#include <vector>
|
|
#include <variant>
|
|
|
|
#include "base/indenter.h"
|
|
#include "base/malloc_arena_pool.h"
|
|
#include "base/scoped_arena_allocator.h"
|
|
#include "builder.h"
|
|
#include "common_compiler_test.h"
|
|
#include "dex/code_item_accessors-inl.h"
|
|
#include "dex/dex_file.h"
|
|
#include "dex/dex_instruction.h"
|
|
#include "dex/standard_dex_file.h"
|
|
#include "driver/dex_compilation_unit.h"
|
|
#include "graph_checker.h"
|
|
#include "gtest/gtest.h"
|
|
#include "handle_scope-inl.h"
|
|
#include "handle_scope.h"
|
|
#include "mirror/class_loader.h"
|
|
#include "mirror/dex_cache.h"
|
|
#include "nodes.h"
|
|
#include "scoped_thread_state_change.h"
|
|
#include "ssa_builder.h"
|
|
#include "ssa_liveness_analysis.h"
|
|
|
|
namespace art {
|
|
|
|
#define NUM_INSTRUCTIONS(...) \
|
|
(sizeof((uint16_t[]) {__VA_ARGS__}) /sizeof(uint16_t))
|
|
|
|
#define N_REGISTERS_CODE_ITEM(NUM_REGS, ...) \
|
|
{ NUM_REGS, 0, 0, 0, 0, 0, NUM_INSTRUCTIONS(__VA_ARGS__), 0, __VA_ARGS__ }
|
|
|
|
#define ZERO_REGISTER_CODE_ITEM(...) N_REGISTERS_CODE_ITEM(0, __VA_ARGS__)
|
|
#define ONE_REGISTER_CODE_ITEM(...) N_REGISTERS_CODE_ITEM(1, __VA_ARGS__)
|
|
#define TWO_REGISTERS_CODE_ITEM(...) N_REGISTERS_CODE_ITEM(2, __VA_ARGS__)
|
|
#define THREE_REGISTERS_CODE_ITEM(...) N_REGISTERS_CODE_ITEM(3, __VA_ARGS__)
|
|
#define FOUR_REGISTERS_CODE_ITEM(...) N_REGISTERS_CODE_ITEM(4, __VA_ARGS__)
|
|
#define FIVE_REGISTERS_CODE_ITEM(...) N_REGISTERS_CODE_ITEM(5, __VA_ARGS__)
|
|
#define SIX_REGISTERS_CODE_ITEM(...) N_REGISTERS_CODE_ITEM(6, __VA_ARGS__)
|
|
|
|
struct InstructionDumper {
|
|
public:
|
|
HInstruction* ins_;
|
|
};
|
|
|
|
inline bool operator==(const InstructionDumper& a, const InstructionDumper& b) {
|
|
return a.ins_ == b.ins_;
|
|
}
|
|
inline bool operator!=(const InstructionDumper& a, const InstructionDumper& b) {
|
|
return !(a == b);
|
|
}
|
|
|
|
inline std::ostream& operator<<(std::ostream& os, const InstructionDumper& id) {
|
|
if (id.ins_ == nullptr) {
|
|
return os << "NULL";
|
|
} else {
|
|
return os << "(" << id.ins_ << "): " << id.ins_->DumpWithArgs();
|
|
}
|
|
}
|
|
|
|
#define EXPECT_INS_EQ(a, b) EXPECT_EQ(InstructionDumper{a}, InstructionDumper{b})
|
|
#define EXPECT_INS_REMOVED(a) EXPECT_TRUE(IsRemoved(a)) << "Not removed: " << (InstructionDumper{a})
|
|
#define EXPECT_INS_RETAINED(a) EXPECT_FALSE(IsRemoved(a)) << "Removed: " << (InstructionDumper{a})
|
|
#define ASSERT_INS_EQ(a, b) ASSERT_EQ(InstructionDumper{a}, InstructionDumper{b})
|
|
#define ASSERT_INS_REMOVED(a) ASSERT_TRUE(IsRemoved(a)) << "Not removed: " << (InstructionDumper{a})
|
|
#define ASSERT_INS_RETAINED(a) ASSERT_FALSE(IsRemoved(a)) << "Removed: " << (InstructionDumper{a})
|
|
|
|
inline LiveInterval* BuildInterval(const size_t ranges[][2],
|
|
size_t number_of_ranges,
|
|
ScopedArenaAllocator* allocator,
|
|
int reg = -1,
|
|
HInstruction* defined_by = nullptr) {
|
|
LiveInterval* interval =
|
|
LiveInterval::MakeInterval(allocator, DataType::Type::kInt32, defined_by);
|
|
if (defined_by != nullptr) {
|
|
defined_by->SetLiveInterval(interval);
|
|
}
|
|
for (size_t i = number_of_ranges; i > 0; --i) {
|
|
interval->AddRange(ranges[i - 1][0], ranges[i - 1][1]);
|
|
}
|
|
interval->SetRegister(reg);
|
|
return interval;
|
|
}
|
|
|
|
inline void RemoveSuspendChecks(HGraph* graph) {
|
|
for (HBasicBlock* block : graph->GetBlocks()) {
|
|
if (block != nullptr) {
|
|
if (block->GetLoopInformation() != nullptr) {
|
|
block->GetLoopInformation()->SetSuspendCheck(nullptr);
|
|
}
|
|
for (HInstructionIterator it(block->GetInstructions()); !it.Done(); it.Advance()) {
|
|
HInstruction* current = it.Current();
|
|
if (current->IsSuspendCheck()) {
|
|
current->GetBlock()->RemoveInstruction(current);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
class ArenaPoolAndAllocator {
|
|
public:
|
|
ArenaPoolAndAllocator()
|
|
: pool_(), allocator_(&pool_), arena_stack_(&pool_), scoped_allocator_(&arena_stack_) { }
|
|
|
|
ArenaAllocator* GetAllocator() { return &allocator_; }
|
|
ArenaStack* GetArenaStack() { return &arena_stack_; }
|
|
ScopedArenaAllocator* GetScopedAllocator() { return &scoped_allocator_; }
|
|
|
|
private:
|
|
MallocArenaPool pool_;
|
|
ArenaAllocator allocator_;
|
|
ArenaStack arena_stack_;
|
|
ScopedArenaAllocator scoped_allocator_;
|
|
};
|
|
|
|
class AdjacencyListGraph {
|
|
public:
|
|
using Edge = std::pair<const std::string_view, const std::string_view>;
|
|
AdjacencyListGraph(
|
|
HGraph* graph,
|
|
ArenaAllocator* alloc,
|
|
const std::string_view entry_name,
|
|
const std::string_view exit_name,
|
|
const std::vector<Edge>& adj) : graph_(graph) {
|
|
auto create_block = [&]() {
|
|
HBasicBlock* blk = new (alloc) HBasicBlock(graph_);
|
|
graph_->AddBlock(blk);
|
|
return blk;
|
|
};
|
|
HBasicBlock* entry = create_block();
|
|
HBasicBlock* exit = create_block();
|
|
graph_->SetEntryBlock(entry);
|
|
graph_->SetExitBlock(exit);
|
|
name_to_block_.Put(entry_name, entry);
|
|
name_to_block_.Put(exit_name, exit);
|
|
for (const auto& [src, dest] : adj) {
|
|
HBasicBlock* src_blk = name_to_block_.GetOrCreate(src, create_block);
|
|
HBasicBlock* dest_blk = name_to_block_.GetOrCreate(dest, create_block);
|
|
src_blk->AddSuccessor(dest_blk);
|
|
}
|
|
graph_->ClearReachabilityInformation();
|
|
graph_->ComputeDominanceInformation();
|
|
graph_->ComputeReachabilityInformation();
|
|
for (auto [name, blk] : name_to_block_) {
|
|
block_to_name_.Put(blk, name);
|
|
}
|
|
}
|
|
|
|
bool HasBlock(const HBasicBlock* blk) const {
|
|
return block_to_name_.find(blk) != block_to_name_.end();
|
|
}
|
|
|
|
std::string_view GetName(const HBasicBlock* blk) const {
|
|
return block_to_name_.Get(blk);
|
|
}
|
|
|
|
HBasicBlock* Get(const std::string_view& sv) const {
|
|
return name_to_block_.Get(sv);
|
|
}
|
|
|
|
AdjacencyListGraph(AdjacencyListGraph&&) = default;
|
|
AdjacencyListGraph(const AdjacencyListGraph&) = default;
|
|
AdjacencyListGraph& operator=(AdjacencyListGraph&&) = default;
|
|
AdjacencyListGraph& operator=(const AdjacencyListGraph&) = default;
|
|
|
|
std::ostream& Dump(std::ostream& os) const {
|
|
struct Namer : public BlockNamer {
|
|
public:
|
|
explicit Namer(const AdjacencyListGraph& alg) : BlockNamer(), alg_(alg) {}
|
|
std::ostream& PrintName(std::ostream& os, HBasicBlock* blk) const override {
|
|
if (alg_.HasBlock(blk)) {
|
|
return os << alg_.GetName(blk) << " (" << blk->GetBlockId() << ")";
|
|
} else {
|
|
return os << "<Unnamed B" << blk->GetBlockId() << ">";
|
|
}
|
|
}
|
|
|
|
const AdjacencyListGraph& alg_;
|
|
};
|
|
Namer namer(*this);
|
|
return graph_->Dump(os, namer);
|
|
}
|
|
|
|
private:
|
|
HGraph* graph_;
|
|
SafeMap<const std::string_view, HBasicBlock*> name_to_block_;
|
|
SafeMap<const HBasicBlock*, const std::string_view> block_to_name_;
|
|
};
|
|
|
|
// Have a separate helper so the OptimizingCFITest can inherit it without causing
|
|
// multiple inheritance errors from having two gtest as a parent twice.
|
|
class OptimizingUnitTestHelper {
|
|
public:
|
|
OptimizingUnitTestHelper()
|
|
: pool_and_allocator_(new ArenaPoolAndAllocator()),
|
|
graph_(nullptr),
|
|
entry_block_(nullptr),
|
|
return_block_(nullptr),
|
|
exit_block_(nullptr) { }
|
|
|
|
ArenaAllocator* GetAllocator() { return pool_and_allocator_->GetAllocator(); }
|
|
ArenaStack* GetArenaStack() { return pool_and_allocator_->GetArenaStack(); }
|
|
ScopedArenaAllocator* GetScopedAllocator() { return pool_and_allocator_->GetScopedAllocator(); }
|
|
|
|
void ResetPoolAndAllocator() {
|
|
pool_and_allocator_.reset(new ArenaPoolAndAllocator());
|
|
}
|
|
|
|
HGraph* CreateGraph(VariableSizedHandleScope* handles = nullptr) {
|
|
ArenaAllocator* const allocator = pool_and_allocator_->GetAllocator();
|
|
|
|
// Reserve a big array of 0s so the dex file constructor can offsets from the header.
|
|
static constexpr size_t kDexDataSize = 4 * KB;
|
|
const uint8_t* dex_data = reinterpret_cast<uint8_t*>(allocator->Alloc(kDexDataSize));
|
|
|
|
// Create the dex file based on the fake data. Call the constructor so that we can use virtual
|
|
// functions. Don't use the arena for the StandardDexFile otherwise the dex location leaks.
|
|
dex_files_.emplace_back(new StandardDexFile(
|
|
dex_data,
|
|
sizeof(StandardDexFile::Header),
|
|
"no_location",
|
|
/*location_checksum*/ 0,
|
|
/*oat_dex_file*/ nullptr,
|
|
/*container*/ nullptr));
|
|
|
|
graph_ = new (allocator) HGraph(
|
|
allocator,
|
|
pool_and_allocator_->GetArenaStack(),
|
|
handles,
|
|
*dex_files_.back(),
|
|
/*method_idx*/-1,
|
|
kRuntimeISA);
|
|
return graph_;
|
|
}
|
|
|
|
// Create a control-flow graph from Dex instructions.
|
|
HGraph* CreateCFG(const std::vector<uint16_t>& data,
|
|
DataType::Type return_type = DataType::Type::kInt32,
|
|
VariableSizedHandleScope* handles = nullptr) {
|
|
HGraph* graph = CreateGraph(handles);
|
|
|
|
// The code item data might not aligned to 4 bytes, copy it to ensure that.
|
|
const size_t code_item_size = data.size() * sizeof(data.front());
|
|
void* aligned_data = GetAllocator()->Alloc(code_item_size);
|
|
memcpy(aligned_data, &data[0], code_item_size);
|
|
CHECK_ALIGNED(aligned_data, StandardDexFile::CodeItem::kAlignment);
|
|
const dex::CodeItem* code_item = reinterpret_cast<const dex::CodeItem*>(aligned_data);
|
|
|
|
{
|
|
const DexCompilationUnit* dex_compilation_unit =
|
|
new (graph->GetAllocator()) DexCompilationUnit(
|
|
/* class_loader= */ Handle<mirror::ClassLoader>(), // Invalid handle.
|
|
/* class_linker= */ nullptr,
|
|
graph->GetDexFile(),
|
|
code_item,
|
|
/* class_def_index= */ DexFile::kDexNoIndex16,
|
|
/* method_idx= */ dex::kDexNoIndex,
|
|
/* access_flags= */ 0u,
|
|
/* verified_method= */ nullptr,
|
|
/* dex_cache= */ Handle<mirror::DexCache>()); // Invalid handle.
|
|
CodeItemDebugInfoAccessor accessor(graph->GetDexFile(), code_item, /*dex_method_idx*/ 0u);
|
|
HGraphBuilder builder(graph, dex_compilation_unit, accessor, return_type);
|
|
bool graph_built = (builder.BuildGraph() == kAnalysisSuccess);
|
|
return graph_built ? graph : nullptr;
|
|
}
|
|
}
|
|
|
|
void InitGraph(VariableSizedHandleScope* handles = nullptr) {
|
|
CreateGraph(handles);
|
|
entry_block_ = AddNewBlock();
|
|
return_block_ = AddNewBlock();
|
|
exit_block_ = AddNewBlock();
|
|
|
|
graph_->SetEntryBlock(entry_block_);
|
|
graph_->SetExitBlock(exit_block_);
|
|
|
|
entry_block_->AddSuccessor(return_block_);
|
|
return_block_->AddSuccessor(exit_block_);
|
|
|
|
return_block_->AddInstruction(new (GetAllocator()) HReturnVoid());
|
|
exit_block_->AddInstruction(new (GetAllocator()) HExit());
|
|
}
|
|
|
|
void AddParameter(HInstruction* parameter) {
|
|
entry_block_->AddInstruction(parameter);
|
|
parameters_.push_back(parameter);
|
|
}
|
|
|
|
HBasicBlock* AddNewBlock() {
|
|
HBasicBlock* block = new (GetAllocator()) HBasicBlock(graph_);
|
|
graph_->AddBlock(block);
|
|
return block;
|
|
}
|
|
|
|
// Run GraphChecker with all checks.
|
|
//
|
|
// Return: the status whether the run is successful.
|
|
bool CheckGraph(HGraph* graph, std::ostream& oss = std::cerr) {
|
|
return CheckGraph(graph, /*check_ref_type_info=*/true, oss);
|
|
}
|
|
|
|
bool CheckGraph(std::ostream& oss = std::cerr) {
|
|
return CheckGraph(graph_, oss);
|
|
}
|
|
|
|
// Run GraphChecker with all checks except reference type information checks.
|
|
//
|
|
// Return: the status whether the run is successful.
|
|
bool CheckGraphSkipRefTypeInfoChecks(HGraph* graph, std::ostream& oss = std::cerr) {
|
|
return CheckGraph(graph, /*check_ref_type_info=*/false, oss);
|
|
}
|
|
|
|
bool CheckGraphSkipRefTypeInfoChecks(std::ostream& oss = std::cerr) {
|
|
return CheckGraphSkipRefTypeInfoChecks(graph_, oss);
|
|
}
|
|
|
|
HEnvironment* ManuallyBuildEnvFor(HInstruction* instruction,
|
|
ArenaVector<HInstruction*>* current_locals) {
|
|
HEnvironment* environment = new (GetAllocator()) HEnvironment(
|
|
(GetAllocator()),
|
|
current_locals->size(),
|
|
graph_->GetArtMethod(),
|
|
instruction->GetDexPc(),
|
|
instruction);
|
|
|
|
environment->CopyFrom(ArrayRef<HInstruction* const>(*current_locals));
|
|
instruction->SetRawEnvironment(environment);
|
|
return environment;
|
|
}
|
|
|
|
void EnsurePredecessorOrder(HBasicBlock* target, std::initializer_list<HBasicBlock*> preds) {
|
|
// Make sure the given preds and block predecessors have the same blocks.
|
|
BitVector bv(preds.size(), false, Allocator::GetMallocAllocator());
|
|
auto preds_and_idx = ZipCount(MakeIterationRange(target->GetPredecessors()));
|
|
bool correct_preds = preds.size() == target->GetPredecessors().size() &&
|
|
std::all_of(preds.begin(), preds.end(), [&](HBasicBlock* pred) {
|
|
return std::any_of(preds_and_idx.begin(),
|
|
preds_and_idx.end(),
|
|
// Make sure every target predecessor is used only
|
|
// once.
|
|
[&](std::pair<HBasicBlock*, uint32_t> cur) {
|
|
if (cur.first == pred && !bv.IsBitSet(cur.second)) {
|
|
bv.SetBit(cur.second);
|
|
return true;
|
|
} else {
|
|
return false;
|
|
}
|
|
});
|
|
}) &&
|
|
bv.NumSetBits() == preds.size();
|
|
auto dump_list = [](auto it) {
|
|
std::ostringstream oss;
|
|
oss << "[";
|
|
bool first = true;
|
|
for (HBasicBlock* b : it) {
|
|
if (!first) {
|
|
oss << ", ";
|
|
}
|
|
first = false;
|
|
oss << b->GetBlockId();
|
|
}
|
|
oss << "]";
|
|
return oss.str();
|
|
};
|
|
ASSERT_TRUE(correct_preds) << "Predecessors of " << target->GetBlockId() << " are "
|
|
<< dump_list(target->GetPredecessors()) << " not "
|
|
<< dump_list(preds);
|
|
if (correct_preds) {
|
|
std::copy(preds.begin(), preds.end(), target->predecessors_.begin());
|
|
}
|
|
}
|
|
|
|
AdjacencyListGraph SetupFromAdjacencyList(const std::string_view entry_name,
|
|
const std::string_view exit_name,
|
|
const std::vector<AdjacencyListGraph::Edge>& adj) {
|
|
return AdjacencyListGraph(graph_, GetAllocator(), entry_name, exit_name, adj);
|
|
}
|
|
|
|
void ManuallyBuildEnvFor(HInstruction* ins, const std::initializer_list<HInstruction*>& env) {
|
|
ArenaVector<HInstruction*> current_locals(env, GetAllocator()->Adapter(kArenaAllocInstruction));
|
|
OptimizingUnitTestHelper::ManuallyBuildEnvFor(ins, ¤t_locals);
|
|
}
|
|
|
|
HLoadClass* MakeClassLoad(std::optional<dex::TypeIndex> ti = std::nullopt,
|
|
std::optional<Handle<mirror::Class>> klass = std::nullopt) {
|
|
return new (GetAllocator()) HLoadClass(graph_->GetCurrentMethod(),
|
|
ti ? *ti : dex::TypeIndex(class_idx_++),
|
|
graph_->GetDexFile(),
|
|
/* klass= */ klass ? *klass : null_klass_,
|
|
/* is_referrers_class= */ false,
|
|
/* dex_pc= */ 0,
|
|
/* needs_access_check= */ false);
|
|
}
|
|
|
|
HNewInstance* MakeNewInstance(HInstruction* cls, uint32_t dex_pc = 0u) {
|
|
EXPECT_TRUE(cls->IsLoadClass() || cls->IsClinitCheck()) << *cls;
|
|
HLoadClass* load =
|
|
cls->IsLoadClass() ? cls->AsLoadClass() : cls->AsClinitCheck()->GetLoadClass();
|
|
return new (GetAllocator()) HNewInstance(cls,
|
|
dex_pc,
|
|
load->GetTypeIndex(),
|
|
graph_->GetDexFile(),
|
|
/* finalizable= */ false,
|
|
QuickEntrypointEnum::kQuickAllocObjectInitialized);
|
|
}
|
|
|
|
HInstanceFieldSet* MakeIFieldSet(HInstruction* inst,
|
|
HInstruction* data,
|
|
MemberOffset off,
|
|
uint32_t dex_pc = 0u) {
|
|
return new (GetAllocator()) HInstanceFieldSet(inst,
|
|
data,
|
|
/* field= */ nullptr,
|
|
/* field_type= */ data->GetType(),
|
|
/* field_offset= */ off,
|
|
/* is_volatile= */ false,
|
|
/* field_idx= */ 0,
|
|
/* declaring_class_def_index= */ 0,
|
|
graph_->GetDexFile(),
|
|
dex_pc);
|
|
}
|
|
|
|
HInstanceFieldGet* MakeIFieldGet(HInstruction* inst,
|
|
DataType::Type type,
|
|
MemberOffset off,
|
|
uint32_t dex_pc = 0u) {
|
|
return new (GetAllocator()) HInstanceFieldGet(inst,
|
|
/* field= */ nullptr,
|
|
/* field_type= */ type,
|
|
/* field_offset= */ off,
|
|
/* is_volatile= */ false,
|
|
/* field_idx= */ 0,
|
|
/* declaring_class_def_index= */ 0,
|
|
graph_->GetDexFile(),
|
|
dex_pc);
|
|
}
|
|
|
|
HInvokeStaticOrDirect* MakeInvoke(DataType::Type return_type,
|
|
const std::vector<HInstruction*>& args) {
|
|
MethodReference method_reference{/* file= */ &graph_->GetDexFile(), /* index= */ method_idx_++};
|
|
HInvokeStaticOrDirect* res = new (GetAllocator())
|
|
HInvokeStaticOrDirect(GetAllocator(),
|
|
args.size(),
|
|
return_type,
|
|
/* dex_pc= */ 0,
|
|
method_reference,
|
|
/* resolved_method= */ nullptr,
|
|
HInvokeStaticOrDirect::DispatchInfo{},
|
|
InvokeType::kStatic,
|
|
/* resolved_method_reference= */ method_reference,
|
|
HInvokeStaticOrDirect::ClinitCheckRequirement::kNone);
|
|
for (auto [ins, idx] : ZipCount(MakeIterationRange(args))) {
|
|
res->SetRawInputAt(idx, ins);
|
|
}
|
|
return res;
|
|
}
|
|
|
|
HPhi* MakePhi(const std::vector<HInstruction*>& ins) {
|
|
EXPECT_GE(ins.size(), 2u) << "Phi requires at least 2 inputs";
|
|
HPhi* phi =
|
|
new (GetAllocator()) HPhi(GetAllocator(), kNoRegNumber, ins.size(), ins[0]->GetType());
|
|
for (auto [i, idx] : ZipCount(MakeIterationRange(ins))) {
|
|
phi->SetRawInputAt(idx, i);
|
|
}
|
|
return phi;
|
|
}
|
|
|
|
void SetupExit(HBasicBlock* exit) {
|
|
exit->AddInstruction(new (GetAllocator()) HExit());
|
|
}
|
|
|
|
dex::TypeIndex DefaultTypeIndexForType(DataType::Type type) {
|
|
switch (type) {
|
|
case DataType::Type::kBool:
|
|
return dex::TypeIndex(1);
|
|
case DataType::Type::kUint8:
|
|
case DataType::Type::kInt8:
|
|
return dex::TypeIndex(2);
|
|
case DataType::Type::kUint16:
|
|
case DataType::Type::kInt16:
|
|
return dex::TypeIndex(3);
|
|
case DataType::Type::kUint32:
|
|
case DataType::Type::kInt32:
|
|
return dex::TypeIndex(4);
|
|
case DataType::Type::kUint64:
|
|
case DataType::Type::kInt64:
|
|
return dex::TypeIndex(5);
|
|
case DataType::Type::kReference:
|
|
return dex::TypeIndex(6);
|
|
case DataType::Type::kFloat32:
|
|
return dex::TypeIndex(7);
|
|
case DataType::Type::kFloat64:
|
|
return dex::TypeIndex(8);
|
|
case DataType::Type::kVoid:
|
|
EXPECT_TRUE(false) << "No type for void!";
|
|
return dex::TypeIndex(1000);
|
|
}
|
|
}
|
|
|
|
// Creates a parameter. The instruction is automatically added to the entry-block
|
|
HParameterValue* MakeParam(DataType::Type type, std::optional<dex::TypeIndex> ti = std::nullopt) {
|
|
HParameterValue* val = new (GetAllocator()) HParameterValue(
|
|
graph_->GetDexFile(), ti ? *ti : DefaultTypeIndexForType(type), param_count_++, type);
|
|
graph_->GetEntryBlock()->AddInstruction(val);
|
|
return val;
|
|
}
|
|
|
|
protected:
|
|
bool CheckGraph(HGraph* graph, bool check_ref_type_info, std::ostream& oss) {
|
|
GraphChecker checker(graph);
|
|
checker.SetRefTypeInfoCheckEnabled(check_ref_type_info);
|
|
checker.Run();
|
|
checker.Dump(oss);
|
|
return checker.IsValid();
|
|
}
|
|
|
|
std::vector<std::unique_ptr<const StandardDexFile>> dex_files_;
|
|
std::unique_ptr<ArenaPoolAndAllocator> pool_and_allocator_;
|
|
|
|
HGraph* graph_;
|
|
HBasicBlock* entry_block_;
|
|
HBasicBlock* return_block_;
|
|
HBasicBlock* exit_block_;
|
|
|
|
std::vector<HInstruction*> parameters_;
|
|
|
|
size_t param_count_ = 0;
|
|
size_t class_idx_ = 42;
|
|
uint32_t method_idx_ = 100;
|
|
|
|
ScopedNullHandle<mirror::Class> null_klass_;
|
|
};
|
|
|
|
class OptimizingUnitTest : public CommonArtTest, public OptimizingUnitTestHelper {};
|
|
|
|
// Naive string diff data type.
|
|
typedef std::list<std::pair<std::string, std::string>> diff_t;
|
|
|
|
// An alias for the empty string used to make it clear that a line is
|
|
// removed in a diff.
|
|
static const std::string removed = ""; // NOLINT [runtime/string] [4]
|
|
|
|
// Naive patch command: apply a diff to a string.
|
|
inline std::string Patch(const std::string& original, const diff_t& diff) {
|
|
std::string result = original;
|
|
for (const auto& p : diff) {
|
|
std::string::size_type pos = result.find(p.first);
|
|
DCHECK_NE(pos, std::string::npos)
|
|
<< "Could not find: \"" << p.first << "\" in \"" << result << "\"";
|
|
result.replace(pos, p.first.size(), p.second);
|
|
}
|
|
return result;
|
|
}
|
|
|
|
// Returns if the instruction is removed from the graph.
|
|
inline bool IsRemoved(HInstruction* instruction) {
|
|
return instruction->GetBlock() == nullptr;
|
|
}
|
|
|
|
inline std::ostream& operator<<(std::ostream& oss, const AdjacencyListGraph& alg) {
|
|
return alg.Dump(oss);
|
|
}
|
|
|
|
class PatternMatchGraphVisitor : public HGraphVisitor {
|
|
private:
|
|
struct HandlerWrapper {
|
|
public:
|
|
virtual ~HandlerWrapper() {}
|
|
virtual void operator()(HInstruction* h) = 0;
|
|
};
|
|
|
|
template <HInstruction::InstructionKind kKind, typename F>
|
|
struct KindWrapper;
|
|
|
|
#define GEN_HANDLER(nm, unused) \
|
|
template <typename F> \
|
|
struct KindWrapper<HInstruction::InstructionKind::k##nm, F> : public HandlerWrapper { \
|
|
public: \
|
|
explicit KindWrapper(F f) : f_(f) {} \
|
|
void operator()(HInstruction* h) override { \
|
|
if constexpr (std::is_invocable_v<F, H##nm*>) { \
|
|
f_(h->As##nm()); \
|
|
} else { \
|
|
LOG(FATAL) << "Incorrect call with " << #nm; \
|
|
} \
|
|
} \
|
|
\
|
|
private: \
|
|
F f_; \
|
|
};
|
|
|
|
FOR_EACH_CONCRETE_INSTRUCTION(GEN_HANDLER)
|
|
#undef GEN_HANDLER
|
|
|
|
template <typename F>
|
|
std::unique_ptr<HandlerWrapper> GetWrapper(HInstruction::InstructionKind kind, F f) {
|
|
switch (kind) {
|
|
#define GEN_GETTER(nm, unused) \
|
|
case HInstruction::InstructionKind::k##nm: \
|
|
return std::unique_ptr<HandlerWrapper>( \
|
|
new KindWrapper<HInstruction::InstructionKind::k##nm, F>(f));
|
|
FOR_EACH_CONCRETE_INSTRUCTION(GEN_GETTER)
|
|
#undef GEN_GETTER
|
|
default:
|
|
LOG(FATAL) << "Unable to handle kind " << kind;
|
|
return nullptr;
|
|
}
|
|
}
|
|
|
|
public:
|
|
template <typename... Inst>
|
|
explicit PatternMatchGraphVisitor(HGraph* graph, Inst... handlers) : HGraphVisitor(graph) {
|
|
FillHandlers(handlers...);
|
|
}
|
|
|
|
void VisitInstruction(HInstruction* instruction) override {
|
|
auto& h = handlers_[instruction->GetKind()];
|
|
if (h.get() != nullptr) {
|
|
(*h)(instruction);
|
|
}
|
|
}
|
|
|
|
private:
|
|
template <typename Func>
|
|
constexpr HInstruction::InstructionKind GetKind() {
|
|
#define CHECK_INST(nm, unused) \
|
|
if constexpr (std::is_invocable_v<Func, H##nm*>) { \
|
|
return HInstruction::InstructionKind::k##nm; \
|
|
}
|
|
FOR_EACH_CONCRETE_INSTRUCTION(CHECK_INST);
|
|
#undef CHECK_INST
|
|
static_assert(!std::is_invocable_v<Func, HInstruction*>,
|
|
"Use on generic HInstruction not allowed");
|
|
#define STATIC_ASSERT_ABSTRACT(nm, unused) && !std::is_invocable_v<Func, H##nm*>
|
|
static_assert(true FOR_EACH_ABSTRACT_INSTRUCTION(STATIC_ASSERT_ABSTRACT),
|
|
"Must not be abstract instruction");
|
|
#undef STATIC_ASSERT_ABSTRACT
|
|
#define STATIC_ASSERT_CONCRETE(nm, unused) || std::is_invocable_v<Func, H##nm*>
|
|
static_assert(false FOR_EACH_CONCRETE_INSTRUCTION(STATIC_ASSERT_CONCRETE),
|
|
"Must be a concrete instruction");
|
|
#undef STATIC_ASSERT_CONCRETE
|
|
return HInstruction::InstructionKind::kLastInstructionKind;
|
|
}
|
|
template <typename First>
|
|
void FillHandlers(First h1) {
|
|
HInstruction::InstructionKind type = GetKind<First>();
|
|
CHECK_NE(type, HInstruction::kLastInstructionKind)
|
|
<< "Unknown instruction kind. Only concrete ones please.";
|
|
handlers_[type] = GetWrapper(type, h1);
|
|
}
|
|
|
|
template <typename First, typename... Inst>
|
|
void FillHandlers(First h1, Inst... handlers) {
|
|
FillHandlers(h1);
|
|
FillHandlers<Inst...>(handlers...);
|
|
}
|
|
|
|
std::array<std::unique_ptr<HandlerWrapper>, HInstruction::InstructionKind::kLastInstructionKind>
|
|
handlers_;
|
|
};
|
|
|
|
template <typename... Target>
|
|
std::tuple<std::vector<Target*>...> FindAllInstructions(
|
|
HGraph* graph,
|
|
std::variant<std::nullopt_t, HBasicBlock*, std::initializer_list<HBasicBlock*>> blks =
|
|
std::nullopt) {
|
|
std::tuple<std::vector<Target*>...> res;
|
|
PatternMatchGraphVisitor vis(
|
|
graph, [&](Target* t) { std::get<std::vector<Target*>>(res).push_back(t); }...);
|
|
|
|
if (std::holds_alternative<std::initializer_list<HBasicBlock*>>(blks)) {
|
|
for (HBasicBlock* blk : std::get<std::initializer_list<HBasicBlock*>>(blks)) {
|
|
vis.VisitBasicBlock(blk);
|
|
}
|
|
} else if (std::holds_alternative<std::nullopt_t>(blks)) {
|
|
vis.VisitInsertionOrder();
|
|
} else {
|
|
vis.VisitBasicBlock(std::get<HBasicBlock*>(blks));
|
|
}
|
|
return res;
|
|
}
|
|
|
|
template <typename... Target>
|
|
std::tuple<Target*...> FindSingleInstructions(
|
|
HGraph* graph,
|
|
std::variant<std::nullopt_t, HBasicBlock*, std::initializer_list<HBasicBlock*>> blks =
|
|
std::nullopt) {
|
|
std::tuple<Target*...> res;
|
|
PatternMatchGraphVisitor vis(graph, [&](Target* t) {
|
|
EXPECT_EQ(std::get<Target*>(res), nullptr)
|
|
<< *std::get<Target*>(res) << " already found but found " << *t << "!";
|
|
std::get<Target*>(res) = t;
|
|
}...);
|
|
if (std::holds_alternative<std::initializer_list<HBasicBlock*>>(blks)) {
|
|
for (HBasicBlock* blk : std::get<std::initializer_list<HBasicBlock*>>(blks)) {
|
|
vis.VisitBasicBlock(blk);
|
|
}
|
|
} else if (std::holds_alternative<std::nullopt_t>(blks)) {
|
|
vis.VisitInsertionOrder();
|
|
} else {
|
|
vis.VisitBasicBlock(std::get<HBasicBlock*>(blks));
|
|
}
|
|
return res;
|
|
}
|
|
|
|
template <typename Target>
|
|
Target* FindSingleInstruction(
|
|
HGraph* graph,
|
|
std::variant<std::nullopt_t, HBasicBlock*, std::initializer_list<HBasicBlock*>> blks =
|
|
std::nullopt) {
|
|
return std::get<Target*>(FindSingleInstructions<Target>(graph, blks));
|
|
}
|
|
|
|
} // namespace art
|
|
|
|
#endif // ART_COMPILER_OPTIMIZING_OPTIMIZING_UNIT_TEST_H_
|