You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
505 lines
17 KiB
505 lines
17 KiB
/*
|
|
* Copyright (C) 2016 The Android Open Source Project
|
|
*
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
#include <iostream>
|
|
#include <type_traits>
|
|
|
|
#include "assembler_arm_vixl.h"
|
|
#include "base/bit_utils.h"
|
|
#include "base/bit_utils_iterator.h"
|
|
#include "entrypoints/quick/quick_entrypoints.h"
|
|
#include "heap_poisoning.h"
|
|
#include "thread.h"
|
|
|
|
using namespace vixl::aarch32; // NOLINT(build/namespaces)
|
|
|
|
namespace art {
|
|
namespace arm {
|
|
|
|
#ifdef ___
|
|
#error "ARM Assembler macro already defined."
|
|
#else
|
|
#define ___ vixl_masm_.
|
|
#endif
|
|
|
|
// Thread register definition.
|
|
extern const vixl32::Register tr(TR);
|
|
// Marking register definition.
|
|
extern const vixl32::Register mr(MR);
|
|
|
|
void ArmVIXLAssembler::FinalizeCode() {
|
|
vixl_masm_.FinalizeCode();
|
|
}
|
|
|
|
size_t ArmVIXLAssembler::CodeSize() const {
|
|
return vixl_masm_.GetSizeOfCodeGenerated();
|
|
}
|
|
|
|
const uint8_t* ArmVIXLAssembler::CodeBufferBaseAddress() const {
|
|
return vixl_masm_.GetBuffer().GetStartAddress<const uint8_t*>();
|
|
}
|
|
|
|
void ArmVIXLAssembler::FinalizeInstructions(const MemoryRegion& region) {
|
|
// Copy the instructions from the buffer.
|
|
MemoryRegion from(vixl_masm_.GetBuffer()->GetStartAddress<void*>(), CodeSize());
|
|
region.CopyFrom(0, from);
|
|
}
|
|
|
|
void ArmVIXLAssembler::PoisonHeapReference(vixl::aarch32::Register reg) {
|
|
// reg = -reg.
|
|
___ Rsb(reg, reg, 0);
|
|
}
|
|
|
|
void ArmVIXLAssembler::UnpoisonHeapReference(vixl::aarch32::Register reg) {
|
|
// reg = -reg.
|
|
___ Rsb(reg, reg, 0);
|
|
}
|
|
|
|
void ArmVIXLAssembler::MaybePoisonHeapReference(vixl32::Register reg) {
|
|
if (kPoisonHeapReferences) {
|
|
PoisonHeapReference(reg);
|
|
}
|
|
}
|
|
|
|
void ArmVIXLAssembler::MaybeUnpoisonHeapReference(vixl32::Register reg) {
|
|
if (kPoisonHeapReferences) {
|
|
UnpoisonHeapReference(reg);
|
|
}
|
|
}
|
|
|
|
void ArmVIXLAssembler::GenerateMarkingRegisterCheck(vixl32::Register temp, int code) {
|
|
// The Marking Register is only used in the Baker read barrier configuration.
|
|
DCHECK(kEmitCompilerReadBarrier);
|
|
DCHECK(kUseBakerReadBarrier);
|
|
|
|
vixl32::Label mr_is_ok;
|
|
|
|
// temp = self.tls32_.is.gc_marking
|
|
___ Ldr(temp, MemOperand(tr, Thread::IsGcMarkingOffset<kArmPointerSize>().Int32Value()));
|
|
// Check that mr == self.tls32_.is.gc_marking.
|
|
___ Cmp(mr, temp);
|
|
___ B(eq, &mr_is_ok, /* is_far_target= */ false);
|
|
___ Bkpt(code);
|
|
___ Bind(&mr_is_ok);
|
|
}
|
|
|
|
void ArmVIXLAssembler::LoadImmediate(vixl32::Register rd, int32_t value) {
|
|
// TODO(VIXL): Implement this optimization in VIXL.
|
|
if (!ShifterOperandCanAlwaysHold(value) && ShifterOperandCanAlwaysHold(~value)) {
|
|
___ Mvn(rd, ~value);
|
|
} else {
|
|
___ Mov(rd, value);
|
|
}
|
|
}
|
|
|
|
bool ArmVIXLAssembler::ShifterOperandCanAlwaysHold(uint32_t immediate) {
|
|
return vixl_masm_.IsModifiedImmediate(immediate);
|
|
}
|
|
|
|
bool ArmVIXLAssembler::ShifterOperandCanHold(Opcode opcode,
|
|
uint32_t immediate,
|
|
vixl::aarch32::FlagsUpdate update_flags) {
|
|
switch (opcode) {
|
|
case ADD:
|
|
case SUB:
|
|
// Less than (or equal to) 12 bits can be done if we don't need to set condition codes.
|
|
if (IsUint<12>(immediate) && update_flags != vixl::aarch32::SetFlags) {
|
|
return true;
|
|
}
|
|
return ShifterOperandCanAlwaysHold(immediate);
|
|
|
|
case MOV:
|
|
// TODO: Support less than or equal to 12bits.
|
|
return ShifterOperandCanAlwaysHold(immediate);
|
|
|
|
case MVN:
|
|
default:
|
|
return ShifterOperandCanAlwaysHold(immediate);
|
|
}
|
|
}
|
|
|
|
bool ArmVIXLAssembler::CanSplitLoadStoreOffset(int32_t allowed_offset_bits,
|
|
int32_t offset,
|
|
/*out*/ int32_t* add_to_base,
|
|
/*out*/ int32_t* offset_for_load_store) {
|
|
int32_t other_bits = offset & ~allowed_offset_bits;
|
|
if (ShifterOperandCanAlwaysHold(other_bits) || ShifterOperandCanAlwaysHold(-other_bits)) {
|
|
*add_to_base = offset & ~allowed_offset_bits;
|
|
*offset_for_load_store = offset & allowed_offset_bits;
|
|
return true;
|
|
}
|
|
return false;
|
|
}
|
|
|
|
int32_t ArmVIXLAssembler::AdjustLoadStoreOffset(int32_t allowed_offset_bits,
|
|
vixl32::Register temp,
|
|
vixl32::Register base,
|
|
int32_t offset) {
|
|
DCHECK_NE(offset & ~allowed_offset_bits, 0);
|
|
int32_t add_to_base, offset_for_load;
|
|
if (CanSplitLoadStoreOffset(allowed_offset_bits, offset, &add_to_base, &offset_for_load)) {
|
|
___ Add(temp, base, add_to_base);
|
|
return offset_for_load;
|
|
} else {
|
|
___ Mov(temp, offset);
|
|
___ Add(temp, temp, base);
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
// TODO(VIXL): Implement this in VIXL.
|
|
int32_t ArmVIXLAssembler::GetAllowedLoadOffsetBits(LoadOperandType type) {
|
|
switch (type) {
|
|
case kLoadSignedByte:
|
|
case kLoadSignedHalfword:
|
|
case kLoadUnsignedHalfword:
|
|
case kLoadUnsignedByte:
|
|
case kLoadWord:
|
|
// We can encode imm12 offset.
|
|
return 0xfff;
|
|
case kLoadSWord:
|
|
case kLoadDWord:
|
|
case kLoadWordPair:
|
|
// We can encode imm8:'00' offset.
|
|
return 0xff << 2;
|
|
default:
|
|
LOG(FATAL) << "UNREACHABLE";
|
|
UNREACHABLE();
|
|
}
|
|
}
|
|
|
|
// TODO(VIXL): Implement this in VIXL.
|
|
int32_t ArmVIXLAssembler::GetAllowedStoreOffsetBits(StoreOperandType type) {
|
|
switch (type) {
|
|
case kStoreHalfword:
|
|
case kStoreByte:
|
|
case kStoreWord:
|
|
// We can encode imm12 offset.
|
|
return 0xfff;
|
|
case kStoreSWord:
|
|
case kStoreDWord:
|
|
case kStoreWordPair:
|
|
// We can encode imm8:'00' offset.
|
|
return 0xff << 2;
|
|
default:
|
|
LOG(FATAL) << "UNREACHABLE";
|
|
UNREACHABLE();
|
|
}
|
|
}
|
|
|
|
// TODO(VIXL): Implement this in VIXL.
|
|
static bool CanHoldLoadOffsetThumb(LoadOperandType type, int offset) {
|
|
switch (type) {
|
|
case kLoadSignedByte:
|
|
case kLoadSignedHalfword:
|
|
case kLoadUnsignedHalfword:
|
|
case kLoadUnsignedByte:
|
|
case kLoadWord:
|
|
return IsAbsoluteUint<12>(offset);
|
|
case kLoadSWord:
|
|
case kLoadDWord:
|
|
return IsAbsoluteUint<10>(offset) && IsAligned<4>(offset); // VFP addressing mode.
|
|
case kLoadWordPair:
|
|
return IsAbsoluteUint<10>(offset) && IsAligned<4>(offset);
|
|
default:
|
|
LOG(FATAL) << "UNREACHABLE";
|
|
UNREACHABLE();
|
|
}
|
|
}
|
|
|
|
// TODO(VIXL): Implement this in VIXL.
|
|
static bool CanHoldStoreOffsetThumb(StoreOperandType type, int offset) {
|
|
switch (type) {
|
|
case kStoreHalfword:
|
|
case kStoreByte:
|
|
case kStoreWord:
|
|
return IsAbsoluteUint<12>(offset);
|
|
case kStoreSWord:
|
|
case kStoreDWord:
|
|
return IsAbsoluteUint<10>(offset) && IsAligned<4>(offset); // VFP addressing mode.
|
|
case kStoreWordPair:
|
|
return IsAbsoluteUint<10>(offset) && IsAligned<4>(offset);
|
|
default:
|
|
LOG(FATAL) << "UNREACHABLE";
|
|
UNREACHABLE();
|
|
}
|
|
}
|
|
|
|
// Implementation note: this method must emit at most one instruction when
|
|
// Address::CanHoldStoreOffsetThumb.
|
|
// TODO(VIXL): Implement AdjustLoadStoreOffset logic in VIXL.
|
|
void ArmVIXLAssembler::StoreToOffset(StoreOperandType type,
|
|
vixl32::Register reg,
|
|
vixl32::Register base,
|
|
int32_t offset) {
|
|
vixl32::Register tmp_reg;
|
|
UseScratchRegisterScope temps(&vixl_masm_);
|
|
|
|
if (!CanHoldStoreOffsetThumb(type, offset)) {
|
|
CHECK_NE(base.GetCode(), kIpCode);
|
|
if ((reg.GetCode() != kIpCode) &&
|
|
(!vixl_masm_.GetScratchRegisterList()->IsEmpty()) &&
|
|
((type != kStoreWordPair) || (reg.GetCode() + 1 != kIpCode))) {
|
|
tmp_reg = temps.Acquire();
|
|
} else {
|
|
// Be careful not to use ip twice (for `reg` (or `reg` + 1 in
|
|
// the case of a word-pair store) and `base`) to build the
|
|
// Address object used by the store instruction(s) below.
|
|
// Instead, save R5 on the stack (or R6 if R5 is already used by
|
|
// `base`), use it as secondary temporary register, and restore
|
|
// it after the store instruction has been emitted.
|
|
tmp_reg = (base.GetCode() != 5) ? r5 : r6;
|
|
___ Push(tmp_reg);
|
|
if (base.GetCode() == kSpCode) {
|
|
offset += kRegisterSize;
|
|
}
|
|
}
|
|
// TODO: Implement indexed store (not available for STRD), inline AdjustLoadStoreOffset()
|
|
// and in the "unsplittable" path get rid of the "add" by using the store indexed instead.
|
|
offset = AdjustLoadStoreOffset(GetAllowedStoreOffsetBits(type), tmp_reg, base, offset);
|
|
base = tmp_reg;
|
|
}
|
|
DCHECK(CanHoldStoreOffsetThumb(type, offset));
|
|
switch (type) {
|
|
case kStoreByte:
|
|
___ Strb(reg, MemOperand(base, offset));
|
|
break;
|
|
case kStoreHalfword:
|
|
___ Strh(reg, MemOperand(base, offset));
|
|
break;
|
|
case kStoreWord:
|
|
___ Str(reg, MemOperand(base, offset));
|
|
break;
|
|
case kStoreWordPair:
|
|
___ Strd(reg, vixl32::Register(reg.GetCode() + 1), MemOperand(base, offset));
|
|
break;
|
|
default:
|
|
LOG(FATAL) << "UNREACHABLE";
|
|
UNREACHABLE();
|
|
}
|
|
if ((tmp_reg.IsValid()) && (tmp_reg.GetCode() != kIpCode)) {
|
|
CHECK(tmp_reg.Is(r5) || tmp_reg.Is(r6)) << tmp_reg;
|
|
___ Pop(tmp_reg);
|
|
}
|
|
}
|
|
|
|
// Implementation note: this method must emit at most one instruction when
|
|
// Address::CanHoldLoadOffsetThumb.
|
|
// TODO(VIXL): Implement AdjustLoadStoreOffset logic in VIXL.
|
|
void ArmVIXLAssembler::LoadFromOffset(LoadOperandType type,
|
|
vixl32::Register dest,
|
|
vixl32::Register base,
|
|
int32_t offset) {
|
|
if (!CanHoldLoadOffsetThumb(type, offset)) {
|
|
CHECK(!base.Is(ip));
|
|
// Inlined AdjustLoadStoreOffset() allows us to pull a few more tricks.
|
|
int32_t allowed_offset_bits = GetAllowedLoadOffsetBits(type);
|
|
DCHECK_NE(offset & ~allowed_offset_bits, 0);
|
|
int32_t add_to_base, offset_for_load;
|
|
if (CanSplitLoadStoreOffset(allowed_offset_bits, offset, &add_to_base, &offset_for_load)) {
|
|
// Use reg for the adjusted base. If it's low reg, we may end up using 16-bit load.
|
|
AddConstant(dest, base, add_to_base);
|
|
base = dest;
|
|
offset = offset_for_load;
|
|
} else {
|
|
UseScratchRegisterScope temps(&vixl_masm_);
|
|
vixl32::Register temp = (dest.Is(base)) ? temps.Acquire() : dest;
|
|
LoadImmediate(temp, offset);
|
|
// TODO: Implement indexed load (not available for LDRD) and use it here to avoid the ADD.
|
|
// Use reg for the adjusted base. If it's low reg, we may end up using 16-bit load.
|
|
___ Add(dest, dest, (dest.Is(base)) ? temp : base);
|
|
base = dest;
|
|
offset = 0;
|
|
}
|
|
}
|
|
|
|
DCHECK(CanHoldLoadOffsetThumb(type, offset));
|
|
switch (type) {
|
|
case kLoadSignedByte:
|
|
___ Ldrsb(dest, MemOperand(base, offset));
|
|
break;
|
|
case kLoadUnsignedByte:
|
|
___ Ldrb(dest, MemOperand(base, offset));
|
|
break;
|
|
case kLoadSignedHalfword:
|
|
___ Ldrsh(dest, MemOperand(base, offset));
|
|
break;
|
|
case kLoadUnsignedHalfword:
|
|
___ Ldrh(dest, MemOperand(base, offset));
|
|
break;
|
|
case kLoadWord:
|
|
CHECK(!dest.IsSP());
|
|
___ Ldr(dest, MemOperand(base, offset));
|
|
break;
|
|
case kLoadWordPair:
|
|
___ Ldrd(dest, vixl32::Register(dest.GetCode() + 1), MemOperand(base, offset));
|
|
break;
|
|
default:
|
|
LOG(FATAL) << "UNREACHABLE";
|
|
UNREACHABLE();
|
|
}
|
|
}
|
|
|
|
void ArmVIXLAssembler::StoreSToOffset(vixl32::SRegister source,
|
|
vixl32::Register base,
|
|
int32_t offset) {
|
|
___ Vstr(source, MemOperand(base, offset));
|
|
}
|
|
|
|
void ArmVIXLAssembler::StoreDToOffset(vixl32::DRegister source,
|
|
vixl32::Register base,
|
|
int32_t offset) {
|
|
___ Vstr(source, MemOperand(base, offset));
|
|
}
|
|
|
|
void ArmVIXLAssembler::LoadSFromOffset(vixl32::SRegister reg,
|
|
vixl32::Register base,
|
|
int32_t offset) {
|
|
___ Vldr(reg, MemOperand(base, offset));
|
|
}
|
|
|
|
void ArmVIXLAssembler::LoadDFromOffset(vixl32::DRegister reg,
|
|
vixl32::Register base,
|
|
int32_t offset) {
|
|
___ Vldr(reg, MemOperand(base, offset));
|
|
}
|
|
|
|
// Prefer Str to Add/Stm in ArmVIXLAssembler::StoreRegisterList and
|
|
// ArmVIXLAssembler::LoadRegisterList where this generates less code (size).
|
|
static constexpr int kRegListThreshold = 4;
|
|
|
|
void ArmVIXLAssembler::StoreRegisterList(RegList regs, size_t stack_offset) {
|
|
int number_of_regs = POPCOUNT(static_cast<uint32_t>(regs));
|
|
if (number_of_regs != 0) {
|
|
if (number_of_regs > kRegListThreshold) {
|
|
UseScratchRegisterScope temps(GetVIXLAssembler());
|
|
vixl32::Register base = sp;
|
|
if (stack_offset != 0) {
|
|
base = temps.Acquire();
|
|
DCHECK_EQ(regs & (1u << base.GetCode()), 0u);
|
|
___ Add(base, sp, Operand::From(stack_offset));
|
|
}
|
|
___ Stm(base, NO_WRITE_BACK, RegisterList(regs));
|
|
} else {
|
|
for (uint32_t i : LowToHighBits(static_cast<uint32_t>(regs))) {
|
|
___ Str(vixl32::Register(i), MemOperand(sp, stack_offset));
|
|
stack_offset += kRegSizeInBytes;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
void ArmVIXLAssembler::LoadRegisterList(RegList regs, size_t stack_offset) {
|
|
int number_of_regs = POPCOUNT(static_cast<uint32_t>(regs));
|
|
if (number_of_regs != 0) {
|
|
if (number_of_regs > kRegListThreshold) {
|
|
UseScratchRegisterScope temps(GetVIXLAssembler());
|
|
vixl32::Register base = sp;
|
|
if (stack_offset != 0) {
|
|
base = temps.Acquire();
|
|
___ Add(base, sp, Operand::From(stack_offset));
|
|
}
|
|
___ Ldm(base, NO_WRITE_BACK, RegisterList(regs));
|
|
} else {
|
|
for (uint32_t i : LowToHighBits(static_cast<uint32_t>(regs))) {
|
|
___ Ldr(vixl32::Register(i), MemOperand(sp, stack_offset));
|
|
stack_offset += kRegSizeInBytes;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
void ArmVIXLAssembler::AddConstant(vixl32::Register rd, int32_t value) {
|
|
AddConstant(rd, rd, value);
|
|
}
|
|
|
|
// TODO(VIXL): think about using adds which updates flags where possible.
|
|
void ArmVIXLAssembler::AddConstant(vixl32::Register rd,
|
|
vixl32::Register rn,
|
|
int32_t value) {
|
|
DCHECK(vixl_masm_.OutsideITBlock());
|
|
// TODO(VIXL): implement this optimization in VIXL.
|
|
if (value == 0) {
|
|
if (!rd.Is(rn)) {
|
|
___ Mov(rd, rn);
|
|
}
|
|
return;
|
|
}
|
|
___ Add(rd, rn, value);
|
|
}
|
|
|
|
// Inside IT block we must use assembler, macroassembler instructions are not permitted.
|
|
void ArmVIXLAssembler::AddConstantInIt(vixl32::Register rd,
|
|
vixl32::Register rn,
|
|
int32_t value,
|
|
vixl32::Condition cond) {
|
|
DCHECK(vixl_masm_.InITBlock());
|
|
if (value == 0) {
|
|
___ mov(cond, rd, rn);
|
|
} else {
|
|
___ add(cond, rd, rn, value);
|
|
}
|
|
}
|
|
|
|
void ArmVIXLMacroAssembler::CompareAndBranchIfZero(vixl32::Register rn,
|
|
vixl32::Label* label,
|
|
bool is_far_target) {
|
|
if (!is_far_target && rn.IsLow() && !label->IsBound()) {
|
|
// In T32, Cbz/Cbnz instructions have following limitations:
|
|
// - There are only 7 bits (i:imm5:0) to encode branch target address (cannot be far target).
|
|
// - Only low registers (i.e R0 .. R7) can be encoded.
|
|
// - Only forward branches (unbound labels) are supported.
|
|
Cbz(rn, label);
|
|
return;
|
|
}
|
|
Cmp(rn, 0);
|
|
B(eq, label, is_far_target);
|
|
}
|
|
|
|
void ArmVIXLMacroAssembler::CompareAndBranchIfNonZero(vixl32::Register rn,
|
|
vixl32::Label* label,
|
|
bool is_far_target) {
|
|
if (!is_far_target && rn.IsLow() && !label->IsBound()) {
|
|
Cbnz(rn, label);
|
|
return;
|
|
}
|
|
Cmp(rn, 0);
|
|
B(ne, label, is_far_target);
|
|
}
|
|
|
|
void ArmVIXLMacroAssembler::B(vixl32::Label* label) {
|
|
if (!label->IsBound()) {
|
|
// Try to use a 16-bit encoding of the B instruction.
|
|
DCHECK(OutsideITBlock());
|
|
BPreferNear(label);
|
|
return;
|
|
}
|
|
MacroAssembler::B(label);
|
|
}
|
|
|
|
void ArmVIXLMacroAssembler::B(vixl32::Condition cond, vixl32::Label* label, bool is_far_target) {
|
|
if (!label->IsBound() && !is_far_target) {
|
|
// Try to use a 16-bit encoding of the B instruction.
|
|
DCHECK(OutsideITBlock());
|
|
BPreferNear(cond, label);
|
|
return;
|
|
}
|
|
MacroAssembler::B(cond, label);
|
|
}
|
|
|
|
} // namespace arm
|
|
} // namespace art
|