/*
* Copyright (C) 2011 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#include "calling_convention.h"
#include "base/logging.h"
#ifdef ART_ENABLE_CODEGEN_arm
#include "jni/quick/arm/calling_convention_arm.h"
#endif
#ifdef ART_ENABLE_CODEGEN_arm64
#include "jni/quick/arm64/calling_convention_arm64.h"
#endif
#ifdef ART_ENABLE_CODEGEN_mips
#include "jni/quick/mips/calling_convention_mips.h"
#endif
#ifdef ART_ENABLE_CODEGEN_mips64
#include "jni/quick/mips64/calling_convention_mips64.h"
#endif
#ifdef ART_ENABLE_CODEGEN_x86
#include "jni/quick/x86/calling_convention_x86.h"
#endif
#ifdef ART_ENABLE_CODEGEN_x86_64
#include "jni/quick/x86_64/calling_convention_x86_64.h"
#endif
namespace art {
// Managed runtime calling convention
std::unique_ptr<ManagedRuntimeCallingConvention> ManagedRuntimeCallingConvention::Create(
ArenaAllocator* arena,
bool is_static,
bool is_synchronized,
const char* shorty,
InstructionSet instruction_set) {
switch (instruction_set) {
#ifdef ART_ENABLE_CODEGEN_arm
case kArm:
case kThumb2:
return std::unique_ptr<ManagedRuntimeCallingConvention>(
new (arena) arm::ArmManagedRuntimeCallingConvention(is_static, is_synchronized, shorty));
#endif
#ifdef ART_ENABLE_CODEGEN_arm64
case kArm64:
return std::unique_ptr<ManagedRuntimeCallingConvention>(
new (arena) arm64::Arm64ManagedRuntimeCallingConvention(
is_static, is_synchronized, shorty));
#endif
#ifdef ART_ENABLE_CODEGEN_mips
case kMips:
return std::unique_ptr<ManagedRuntimeCallingConvention>(
new (arena) mips::MipsManagedRuntimeCallingConvention(
is_static, is_synchronized, shorty));
#endif
#ifdef ART_ENABLE_CODEGEN_mips64
case kMips64:
return std::unique_ptr<ManagedRuntimeCallingConvention>(
new (arena) mips64::Mips64ManagedRuntimeCallingConvention(
is_static, is_synchronized, shorty));
#endif
#ifdef ART_ENABLE_CODEGEN_x86
case kX86:
return std::unique_ptr<ManagedRuntimeCallingConvention>(
new (arena) x86::X86ManagedRuntimeCallingConvention(is_static, is_synchronized, shorty));
#endif
#ifdef ART_ENABLE_CODEGEN_x86_64
case kX86_64:
return std::unique_ptr<ManagedRuntimeCallingConvention>(
new (arena) x86_64::X86_64ManagedRuntimeCallingConvention(
is_static, is_synchronized, shorty));
#endif
default:
LOG(FATAL) << "Unknown InstructionSet: " << instruction_set;
UNREACHABLE();
}
}
bool ManagedRuntimeCallingConvention::HasNext() {
return itr_args_ < NumArgs();
}
void ManagedRuntimeCallingConvention::Next() {
CHECK(HasNext());
if (IsCurrentArgExplicit() && // don't query parameter type of implicit args
IsParamALongOrDouble(itr_args_)) {
itr_longs_and_doubles_++;
itr_slots_++;
}
if (IsParamAFloatOrDouble(itr_args_)) {
itr_float_and_doubles_++;
}
if (IsCurrentParamAReference()) {
itr_refs_++;
}
itr_args_++;
itr_slots_++;
}
bool ManagedRuntimeCallingConvention::IsCurrentArgExplicit() {
// Static methods have no implicit arguments, others implicitly pass this
return IsStatic() || (itr_args_ != 0);
}
bool ManagedRuntimeCallingConvention::IsCurrentArgPossiblyNull() {
return IsCurrentArgExplicit(); // any user parameter may be null
}
size_t ManagedRuntimeCallingConvention::CurrentParamSize() {
return ParamSize(itr_args_);
}
bool ManagedRuntimeCallingConvention::IsCurrentParamAReference() {
return IsParamAReference(itr_args_);
}
bool ManagedRuntimeCallingConvention::IsCurrentParamAFloatOrDouble() {
return IsParamAFloatOrDouble(itr_args_);
}
bool ManagedRuntimeCallingConvention::IsCurrentParamADouble() {
return IsParamADouble(itr_args_);
}
bool ManagedRuntimeCallingConvention::IsCurrentParamALong() {
return IsParamALong(itr_args_);
}
// JNI calling convention
std::unique_ptr<JniCallingConvention> JniCallingConvention::Create(ArenaAllocator* arena,
bool is_static,
bool is_synchronized,
bool is_critical_native,
const char* shorty,
InstructionSet instruction_set) {
switch (instruction_set) {
#ifdef ART_ENABLE_CODEGEN_arm
case kArm:
case kThumb2:
return std::unique_ptr<JniCallingConvention>(
new (arena) arm::ArmJniCallingConvention(is_static,
is_synchronized,
is_critical_native,
shorty));
#endif
#ifdef ART_ENABLE_CODEGEN_arm64
case kArm64:
return std::unique_ptr<JniCallingConvention>(
new (arena) arm64::Arm64JniCallingConvention(is_static,
is_synchronized,
is_critical_native,
shorty));
#endif
#ifdef ART_ENABLE_CODEGEN_mips
case kMips:
return std::unique_ptr<JniCallingConvention>(
new (arena) mips::MipsJniCallingConvention(is_static,
is_synchronized,
is_critical_native,
shorty));
#endif
#ifdef ART_ENABLE_CODEGEN_mips64
case kMips64:
return std::unique_ptr<JniCallingConvention>(
new (arena) mips64::Mips64JniCallingConvention(is_static,
is_synchronized,
is_critical_native,
shorty));
#endif
#ifdef ART_ENABLE_CODEGEN_x86
case kX86:
return std::unique_ptr<JniCallingConvention>(
new (arena) x86::X86JniCallingConvention(is_static,
is_synchronized,
is_critical_native,
shorty));
#endif
#ifdef ART_ENABLE_CODEGEN_x86_64
case kX86_64:
return std::unique_ptr<JniCallingConvention>(
new (arena) x86_64::X86_64JniCallingConvention(is_static,
is_synchronized,
is_critical_native,
shorty));
#endif
default:
LOG(FATAL) << "Unknown InstructionSet: " << instruction_set;
UNREACHABLE();
}
}
size_t JniCallingConvention::ReferenceCount() const {
return NumReferenceArgs() + (IsStatic() ? 1 : 0);
}
FrameOffset JniCallingConvention::SavedLocalReferenceCookieOffset() const {
size_t references_size = handle_scope_pointer_size_ * ReferenceCount(); // size excluding header
return FrameOffset(HandleReferencesOffset().Int32Value() + references_size);
}
FrameOffset JniCallingConvention::ReturnValueSaveLocation() const {
if (LIKELY(HasHandleScope())) {
// Initial offset already includes the displacement.
// -- Remove the additional local reference cookie offset if we don't have a handle scope.
const size_t saved_local_reference_cookie_offset =
SavedLocalReferenceCookieOffset().Int32Value();
// Segment state is 4 bytes long
const size_t segment_state_size = 4;
return FrameOffset(saved_local_reference_cookie_offset + segment_state_size);
} else {
// Include only the initial Method* as part of the offset.
CHECK_LT(displacement_.SizeValue(),
static_cast<size_t>(std::numeric_limits<int32_t>::max()));
return FrameOffset(displacement_.Int32Value() + static_cast<size_t>(frame_pointer_size_));
}
}
bool JniCallingConvention::HasNext() {
if (IsCurrentArgExtraForJni()) {
return true;
} else {
unsigned int arg_pos = GetIteratorPositionWithinShorty();
return arg_pos < NumArgs();
}
}
void JniCallingConvention::Next() {
CHECK(HasNext());
if (IsCurrentParamALong() || IsCurrentParamADouble()) {
itr_longs_and_doubles_++;
itr_slots_++;
}
if (IsCurrentParamAFloatOrDouble()) {
itr_float_and_doubles_++;
}
if (IsCurrentParamAReference()) {
itr_refs_++;
}
// This default/fallthrough case also covers the extra JNIEnv* argument,
// as well as any other single-slot primitives.
itr_args_++;
itr_slots_++;
}
bool JniCallingConvention::IsCurrentParamAReference() {
bool return_value;
if (SwitchExtraJniArguments(itr_args_,
false, // JNIEnv*
true, // jobject or jclass
/* out parameters */
&return_value)) {
return return_value;
} else {
int arg_pos = GetIteratorPositionWithinShorty();
return IsParamAReference(arg_pos);
}
}
bool JniCallingConvention::IsCurrentParamJniEnv() {
if (UNLIKELY(!HasJniEnv())) {
return false;
}
return (itr_args_ == kJniEnv);
}
bool JniCallingConvention::IsCurrentParamAFloatOrDouble() {
bool return_value;
if (SwitchExtraJniArguments(itr_args_,
false, // jnienv*
false, // jobject or jclass
/* out parameters */
&return_value)) {
return return_value;
} else {
int arg_pos = GetIteratorPositionWithinShorty();
return IsParamAFloatOrDouble(arg_pos);
}
}
bool JniCallingConvention::IsCurrentParamADouble() {
bool return_value;
if (SwitchExtraJniArguments(itr_args_,
false, // jnienv*
false, // jobject or jclass
/* out parameters */
&return_value)) {
return return_value;
} else {
int arg_pos = GetIteratorPositionWithinShorty();
return IsParamADouble(arg_pos);
}
}
bool JniCallingConvention::IsCurrentParamALong() {
bool return_value;
if (SwitchExtraJniArguments(itr_args_,
false, // jnienv*
false, // jobject or jclass
/* out parameters */
&return_value)) {
return return_value;
} else {
int arg_pos = GetIteratorPositionWithinShorty();
return IsParamALong(arg_pos);
}
}
// Return position of handle scope entry holding reference at the current iterator
// position
FrameOffset JniCallingConvention::CurrentParamHandleScopeEntryOffset() {
CHECK(IsCurrentParamAReference());
CHECK_LT(HandleScopeLinkOffset(), HandleScopeNumRefsOffset());
int result = HandleReferencesOffset().Int32Value() + itr_refs_ * handle_scope_pointer_size_;
CHECK_GT(result, HandleScopeNumRefsOffset().Int32Value());
return FrameOffset(result);
}
size_t JniCallingConvention::CurrentParamSize() const {
if (IsCurrentArgExtraForJni()) {
return static_cast<size_t>(frame_pointer_size_); // JNIEnv or jobject/jclass
} else {
int arg_pos = GetIteratorPositionWithinShorty();
return ParamSize(arg_pos);
}
}
size_t JniCallingConvention::NumberOfExtraArgumentsForJni() const {
if (LIKELY(HasExtraArgumentsForJni())) {
// The first argument is the JNIEnv*.
// Static methods have an extra argument which is the jclass.
return IsStatic() ? 2 : 1;
} else {
// Critical natives exclude the JNIEnv and the jclass/this parameters.
return 0;
}
}
bool JniCallingConvention::HasHandleScope() const {
// Exclude HandleScope for @CriticalNative methods for optimization speed.
return is_critical_native_ == false;
}
bool JniCallingConvention::HasLocalReferenceSegmentState() const {
// Exclude local reference segment states for @CriticalNative methods for optimization speed.
return is_critical_native_ == false;
}
bool JniCallingConvention::HasJniEnv() const {
// Exclude "JNIEnv*" parameter for @CriticalNative methods.
return HasExtraArgumentsForJni();
}
bool JniCallingConvention::HasSelfClass() const {
if (!IsStatic()) {
// Virtual functions: There is never an implicit jclass parameter.
return false;
} else {
// Static functions: There is an implicit jclass parameter unless it's @CriticalNative.
return HasExtraArgumentsForJni();
}
}
bool JniCallingConvention::HasExtraArgumentsForJni() const {
// @CriticalNative jni implementations exclude both JNIEnv* and the jclass/jobject parameters.
return is_critical_native_ == false;
}
unsigned int JniCallingConvention::GetIteratorPositionWithinShorty() const {
// We need to subtract out the extra JNI arguments if we want to use this iterator position
// with the inherited CallingConvention member functions, which rely on scanning the shorty.
// Note that our shorty does *not* include the JNIEnv, jclass/jobject parameters.
DCHECK_GE(itr_args_, NumberOfExtraArgumentsForJni());
return itr_args_ - NumberOfExtraArgumentsForJni();
}
bool JniCallingConvention::IsCurrentArgExtraForJni() const {
if (UNLIKELY(!HasExtraArgumentsForJni())) {
return false; // If there are no extra args, we can never be an extra.
}
// Only parameters kJniEnv and kObjectOrClass are considered extra.
return itr_args_ <= kObjectOrClass;
}
bool JniCallingConvention::SwitchExtraJniArguments(size_t switch_value,
bool case_jni_env,
bool case_object_or_class,
/* out parameters */
bool* return_value) const {
DCHECK(return_value != nullptr);
if (UNLIKELY(!HasExtraArgumentsForJni())) {
return false;
}
switch (switch_value) {
case kJniEnv:
*return_value = case_jni_env;
return true;
case kObjectOrClass:
*return_value = case_object_or_class;
return true;
default:
return false;
}
}
} // namespace art