gcc/libjava/verify.cc

3311 lines
84 KiB
C++
Raw Normal View History

// verify.cc - verify bytecode
/* Copyright (C) 2001, 2002, 2003 Free Software Foundation
This file is part of libgcj.
This software is copyrighted work licensed under the terms of the
Libgcj License. Please consult the file "LIBGCJ_LICENSE" for
details. */
// Written by Tom Tromey <tromey@redhat.com>
// Define VERIFY_DEBUG to enable debugging output.
#include <config.h>
#include <jvm.h>
#include <gcj/cni.h>
#include <java-insns.h>
#include <java-interp.h>
#ifdef INTERPRETER
#include <java/lang/Class.h>
#include <java/lang/VerifyError.h>
#include <java/lang/Throwable.h>
#include <java/lang/reflect/Modifier.h>
#include <java/lang/StringBuffer.h>
#ifdef VERIFY_DEBUG
#include <stdio.h>
#endif /* VERIFY_DEBUG */
static void debug_print (const char *fmt, ...)
__attribute__ ((format (printf, 1, 2)));
static inline void
debug_print (const char *fmt, ...)
{
#ifdef VERIFY_DEBUG
va_list ap;
va_start (ap, fmt);
vfprintf (stderr, fmt, ap);
va_end (ap);
#endif /* VERIFY_DEBUG */
}
class _Jv_BytecodeVerifier
{
private:
static const int FLAG_INSN_START = 1;
static const int FLAG_BRANCH_TARGET = 2;
struct state;
struct type;
struct subr_info;
struct subr_entry_info;
struct linked_utf8;
struct ref_intersection;
// The current PC.
int PC;
// The PC corresponding to the start of the current instruction.
int start_PC;
// The current state of the stack, locals, etc.
state *current_state;
// We store the state at branch targets, for merging. This holds
// such states.
state **states;
// We keep a linked list of all the PCs which we must reverify.
// The link is done using the PC values. This is the head of the
// list.
int next_verify_pc;
// We keep some flags for each instruction. The values are the
// FLAG_* constants defined above.
char *flags;
// We need to keep track of which instructions can call a given
// subroutine. FIXME: this is inefficient. We keep a linked list
// of all calling `jsr's at at each jsr target.
subr_info **jsr_ptrs;
// We keep a linked list of entries which map each `ret' instruction
// to its unique subroutine entry point. We expect that there won't
// be many `ret' instructions, so a linked list is ok.
subr_entry_info *entry_points;
// The bytecode itself.
unsigned char *bytecode;
// The exceptions.
_Jv_InterpException *exception;
// Defining class.
jclass current_class;
// This method.
_Jv_InterpMethod *current_method;
// A linked list of utf8 objects we allocate. This is really ugly,
// but without this our utf8 objects would be collected.
linked_utf8 *utf8_list;
// A linked list of all ref_intersection objects we allocate.
ref_intersection *isect_list;
struct linked_utf8
{
_Jv_Utf8Const *val;
linked_utf8 *next;
};
_Jv_Utf8Const *make_utf8_const (char *s, int len)
{
_Jv_Utf8Const *val = _Jv_makeUtf8Const (s, len);
_Jv_Utf8Const *r = (_Jv_Utf8Const *) _Jv_Malloc (sizeof (_Jv_Utf8Const)
+ val->length
+ 1);
r->length = val->length;
r->hash = val->hash;
memcpy (r->data, val->data, val->length + 1);
linked_utf8 *lu = (linked_utf8 *) _Jv_Malloc (sizeof (linked_utf8));
lu->val = r;
lu->next = utf8_list;
utf8_list = lu;
return r;
}
__attribute__ ((__noreturn__)) void verify_fail (char *s, jint pc = -1)
{
using namespace java::lang;
StringBuffer *buf = new StringBuffer ();
buf->append (JvNewStringLatin1 ("verification failed"));
if (pc == -1)
pc = start_PC;
if (pc != -1)
{
buf->append (JvNewStringLatin1 (" at PC "));
buf->append (pc);
}
_Jv_InterpMethod *method = current_method;
buf->append (JvNewStringLatin1 (" in "));
buf->append (current_class->getName());
buf->append ((jchar) ':');
buf->append (JvNewStringUTF (method->get_method()->name->data));
buf->append ((jchar) '(');
buf->append (JvNewStringUTF (method->get_method()->signature->data));
buf->append ((jchar) ')');
buf->append (JvNewStringLatin1 (": "));
buf->append (JvNewStringLatin1 (s));
throw new java::lang::VerifyError (buf->toString ());
}
// This enum holds a list of tags for all the different types we
// need to handle. Reference types are treated specially by the
// type class.
enum type_val
{
void_type,
// The values for primitive types are chosen to correspond to values
// specified to newarray.
boolean_type = 4,
char_type = 5,
float_type = 6,
double_type = 7,
byte_type = 8,
short_type = 9,
int_type = 10,
long_type = 11,
// Used when overwriting second word of a double or long in the
// local variables. Also used after merging local variable states
// to indicate an unusable value.
unsuitable_type,
return_address_type,
continuation_type,
// There is an obscure special case which requires us to note when
// a local variable has not been used by a subroutine. See
// push_jump_merge for more information.
unused_by_subroutine_type,
// Everything after `reference_type' must be a reference type.
reference_type,
null_type,
uninitialized_reference_type
};
// This represents a merged class type. Some verifiers (including
// earlier versions of this one) will compute the intersection of
// two class types when merging states. However, this loses
// critical information about interfaces implemented by the various
// classes. So instead we keep track of all the actual classes that
// have been merged.
struct ref_intersection
{
// Whether or not this type has been resolved.
bool is_resolved;
// Actual type data.
union
{
// For a resolved reference type, this is a pointer to the class.
jclass klass;
// For other reference types, this it the name of the class.
_Jv_Utf8Const *name;
} data;
// Link to the next reference in the intersection.
ref_intersection *ref_next;
// This is used to keep track of all the allocated
// ref_intersection objects, so we can free them.
// FIXME: we should allocate these in chunks.
ref_intersection *alloc_next;
ref_intersection (jclass klass, _Jv_BytecodeVerifier *verifier)
: ref_next (NULL)
{
is_resolved = true;
data.klass = klass;
alloc_next = verifier->isect_list;
verifier->isect_list = this;
}
ref_intersection (_Jv_Utf8Const *name, _Jv_BytecodeVerifier *verifier)
: ref_next (NULL)
{
is_resolved = false;
data.name = name;
alloc_next = verifier->isect_list;
verifier->isect_list = this;
}
ref_intersection (ref_intersection *dup, ref_intersection *tail,
_Jv_BytecodeVerifier *verifier)
: ref_next (tail)
{
is_resolved = dup->is_resolved;
data = dup->data;
alloc_next = verifier->isect_list;
verifier->isect_list = this;
}
bool equals (ref_intersection *other, _Jv_BytecodeVerifier *verifier)
{
if (! is_resolved && ! other->is_resolved
&& _Jv_equalUtf8Consts (data.name, other->data.name))
return true;
if (! is_resolved)
resolve (verifier);
if (! other->is_resolved)
other->resolve (verifier);
return data.klass == other->data.klass;
}
// Merge THIS type into OTHER, returning the result. This will
// return OTHER if all the classes in THIS already appear in
// OTHER.
ref_intersection *merge (ref_intersection *other,
_Jv_BytecodeVerifier *verifier)
{
ref_intersection *tail = other;
for (ref_intersection *self = this; self != NULL; self = self->ref_next)
{
bool add = true;
for (ref_intersection *iter = other; iter != NULL;
iter = iter->ref_next)
{
if (iter->equals (self, verifier))
{
add = false;
break;
}
}
if (add)
tail = new ref_intersection (self, tail, verifier);
}
return tail;
}
void resolve (_Jv_BytecodeVerifier *verifier)
{
if (is_resolved)
return;
using namespace java::lang;
java::lang::ClassLoader *loader
= verifier->current_class->getClassLoaderInternal();
// We might see either kind of name. Sigh.
if (data.name->data[0] == 'L'
&& data.name->data[data.name->length - 1] == ';')
data.klass = _Jv_FindClassFromSignature (data.name->data, loader);
else
data.klass = Class::forName (_Jv_NewStringUtf8Const (data.name),
false, loader);
is_resolved = true;
}
// See if an object of type OTHER can be assigned to an object of
// type *THIS. This might resolve classes in one chain or the
// other.
bool compatible (ref_intersection *other,
_Jv_BytecodeVerifier *verifier)
{
ref_intersection *self = this;
for (; self != NULL; self = self->ref_next)
{
ref_intersection *other_iter = other;
for (; other_iter != NULL; other_iter = other_iter->ref_next)
{
// Avoid resolving if possible.
if (! self->is_resolved
&& ! other_iter->is_resolved
&& _Jv_equalUtf8Consts (self->data.name,
other_iter->data.name))
continue;
if (! self->is_resolved)
self->resolve(verifier);
if (! other_iter->is_resolved)
other_iter->resolve(verifier);
if (! is_assignable_from_slow (self->data.klass,
other_iter->data.klass))
return false;
}
}
return true;
}
bool isarray ()
{
// assert (ref_next == NULL);
if (is_resolved)
return data.klass->isArray ();
else
return data.name->data[0] == '[';
}
bool isinterface (_Jv_BytecodeVerifier *verifier)
{
// assert (ref_next == NULL);
if (! is_resolved)
resolve (verifier);
return data.klass->isInterface ();
}
bool isabstract (_Jv_BytecodeVerifier *verifier)
{
// assert (ref_next == NULL);
if (! is_resolved)
resolve (verifier);
using namespace java::lang::reflect;
return Modifier::isAbstract (data.klass->getModifiers ());
}
jclass getclass (_Jv_BytecodeVerifier *verifier)
{
if (! is_resolved)
resolve (verifier);
return data.klass;
}
int count_dimensions ()
{
int ndims = 0;
if (is_resolved)
{
jclass k = data.klass;
while (k->isArray ())
{
k = k->getComponentType ();
++ndims;
}
}
else
{
char *p = data.name->data;
while (*p++ == '[')
++ndims;
}
return ndims;
}
void *operator new (size_t bytes)
{
return _Jv_Malloc (bytes);
}
void operator delete (void *mem)
{
_Jv_Free (mem);
}
};
// Return the type_val corresponding to a primitive signature
// character. For instance `I' returns `int.class'.
type_val get_type_val_for_signature (jchar sig)
{
type_val rt;
switch (sig)
{
case 'Z':
rt = boolean_type;
break;
case 'B':
rt = byte_type;
break;
case 'C':
rt = char_type;
break;
case 'S':
rt = short_type;
break;
case 'I':
rt = int_type;
break;
case 'J':
rt = long_type;
break;
case 'F':
rt = float_type;
break;
case 'D':
rt = double_type;
break;
case 'V':
rt = void_type;
break;
default:
verify_fail ("invalid signature");
}
return rt;
}
// Return the type_val corresponding to a primitive class.
type_val get_type_val_for_signature (jclass k)
{
return get_type_val_for_signature ((jchar) k->method_count);
}
// This is like _Jv_IsAssignableFrom, but it works even if SOURCE or
// TARGET haven't been prepared.
static bool is_assignable_from_slow (jclass target, jclass source)
{
// First, strip arrays.
while (target->isArray ())
{
// If target is array, source must be as well.
if (! source->isArray ())
return false;
target = target->getComponentType ();
source = source->getComponentType ();
}
// Quick success.
if (target == &java::lang::Object::class$)
return true;
do
{
if (source == target)
return true;
if (target->isPrimitive () || source->isPrimitive ())
return false;
if (target->isInterface ())
{
for (int i = 0; i < source->interface_count; ++i)
{
// We use a recursive call because we also need to
// check superinterfaces.
if (is_assignable_from_slow (target, source->interfaces[i]))
return true;
}
}
source = source->getSuperclass ();
}
while (source != NULL);
return false;
}
// This is used to keep track of which `jsr's correspond to a given
// jsr target.
struct subr_info
{
// PC of the instruction just after the jsr.
int pc;
// Link.
subr_info *next;
};
// This is used to keep track of which subroutine entry point
// corresponds to which `ret' instruction.
struct subr_entry_info
{
// PC of the subroutine entry point.
int pc;
// PC of the `ret' instruction.
int ret_pc;
// Link.
subr_entry_info *next;
};
// The `type' class is used to represent a single type in the
// verifier.
struct type
{
// The type key.
type_val key;
// For reference types, the representation of the type.
ref_intersection *klass;
// This is used when constructing a new object. It is the PC of the
// `new' instruction which created the object. We use the special
// value -2 to mean that this is uninitialized, and the special
// value -1 for the case where the current method is itself the
// <init> method.
int pc;
static const int UNINIT = -2;
static const int SELF = -1;
// Basic constructor.
type ()
{
key = unsuitable_type;
klass = NULL;
pc = UNINIT;
}
// Make a new instance given the type tag. We assume a generic
// `reference_type' means Object.
type (type_val k)
{
key = k;
// For reference_type, if KLASS==NULL then that means we are
// looking for a generic object of any kind, including an
// uninitialized reference.
klass = NULL;
pc = UNINIT;
}
// Make a new instance given a class.
type (jclass k, _Jv_BytecodeVerifier *verifier)
{
key = reference_type;
klass = new ref_intersection (k, verifier);
pc = UNINIT;
}
// Make a new instance given the name of a class.
type (_Jv_Utf8Const *n, _Jv_BytecodeVerifier *verifier)
{
key = reference_type;
klass = new ref_intersection (n, verifier);
pc = UNINIT;
}
// Copy constructor.
type (const type &t)
{
key = t.key;
klass = t.klass;
pc = t.pc;
}
// These operators are required because libgcj can't link in
// -lstdc++.
void *operator new[] (size_t bytes)
{
return _Jv_Malloc (bytes);
}
void operator delete[] (void *mem)
{
_Jv_Free (mem);
}
type& operator= (type_val k)
{
key = k;
klass = NULL;
pc = UNINIT;
return *this;
}
type& operator= (const type& t)
{
key = t.key;
klass = t.klass;
pc = t.pc;
return *this;
}
// Promote a numeric type.
type &promote ()
{
if (key == boolean_type || key == char_type
|| key == byte_type || key == short_type)
key = int_type;
return *this;
}
// Mark this type as the uninitialized result of `new'.
void set_uninitialized (int npc, _Jv_BytecodeVerifier *verifier)
{
if (key == reference_type)
key = uninitialized_reference_type;
else
verifier->verify_fail ("internal error in type::uninitialized");
pc = npc;
}
// Mark this type as now initialized.
void set_initialized (int npc)
{
if (npc != UNINIT && pc == npc && key == uninitialized_reference_type)
{
key = reference_type;
pc = UNINIT;
}
}
// Return true if an object of type K can be assigned to a variable
// of type *THIS. Handle various special cases too. Might modify
// *THIS or K. Note however that this does not perform numeric
// promotion.
bool compatible (type &k, _Jv_BytecodeVerifier *verifier)
{
// Any type is compatible with the unsuitable type.
if (key == unsuitable_type)
return true;
if (key < reference_type || k.key < reference_type)
return key == k.key;
// The `null' type is convertible to any initialized reference
// type.
if (key == null_type)
return k.key != uninitialized_reference_type;
if (k.key == null_type)
return key != uninitialized_reference_type;
// A special case for a generic reference.
if (klass == NULL)
return true;
if (k.klass == NULL)
verifier->verify_fail ("programmer error in type::compatible");
// An initialized type and an uninitialized type are not
// compatible.
if (isinitialized () != k.isinitialized ())
return false;
// Two uninitialized objects are compatible if either:
// * The PCs are identical, or
// * One PC is UNINIT.
if (! isinitialized ())
{
if (pc != k.pc && pc != UNINIT && k.pc != UNINIT)
return false;
}
return klass->compatible(k.klass, verifier);
}
bool isvoid () const
{
return key == void_type;
}
bool iswide () const
{
return key == long_type || key == double_type;
}
// Return number of stack or local variable slots taken by this
// type.
int depth () const
{
return iswide () ? 2 : 1;
}
bool isarray () const
{
// We treat null_type as not an array. This is ok based on the
// current uses of this method.
if (key == reference_type)
return klass->isarray ();
return false;
}
bool isnull () const
{
return key == null_type;
}
bool isinterface (_Jv_BytecodeVerifier *verifier)
{
if (key != reference_type)
return false;
return klass->isinterface (verifier);
}
bool isabstract (_Jv_BytecodeVerifier *verifier)
{
if (key != reference_type)
return false;
return klass->isabstract (verifier);
}
// Return the element type of an array.
type element_type (_Jv_BytecodeVerifier *verifier)
{
if (key != reference_type)
verifier->verify_fail ("programmer error in type::element_type()", -1);
jclass k = klass->getclass (verifier)->getComponentType ();
if (k->isPrimitive ())
return type (verifier->get_type_val_for_signature (k));
return type (k, verifier);
}
// Return the array type corresponding to an initialized
// reference. We could expand this to work for other kinds of
// types, but currently we don't need to.
type to_array (_Jv_BytecodeVerifier *verifier)
{
if (key != reference_type)
verifier->verify_fail ("internal error in type::to_array()");
jclass k = klass->getclass (verifier);
return type (_Jv_GetArrayClass (k, k->getClassLoaderInternal()),
verifier);
}
bool isreference () const
{
return key >= reference_type;
}
int get_pc () const
{
return pc;
}
bool isinitialized () const
{
return key == reference_type || key == null_type;
}
bool isresolved () const
{
return (key == reference_type
|| key == null_type
|| key == uninitialized_reference_type);
}
void verify_dimensions (int ndims, _Jv_BytecodeVerifier *verifier)
{
// The way this is written, we don't need to check isarray().
if (key != reference_type)
verifier->verify_fail ("internal error in verify_dimensions: not a reference type");
if (klass->count_dimensions () < ndims)
verifier->verify_fail ("array type has fewer dimensions than required");
}
// Merge OLD_TYPE into this. On error throw exception.
bool merge (type& old_type, bool local_semantics,
_Jv_BytecodeVerifier *verifier)
{
bool changed = false;
bool refo = old_type.isreference ();
bool refn = isreference ();
if (refo && refn)
{
if (old_type.key == null_type)
;
else if (key == null_type)
{
*this = old_type;
changed = true;
}
else if (isinitialized () != old_type.isinitialized ())
verifier->verify_fail ("merging initialized and uninitialized types");
else
{
if (! isinitialized ())
{
if (pc == UNINIT)
pc = old_type.pc;
else if (old_type.pc == UNINIT)
;
else if (pc != old_type.pc)
verifier->verify_fail ("merging different uninitialized types");
}
ref_intersection *merged = old_type.klass->merge (klass,
verifier);
if (merged != klass)
{
klass = merged;
changed = true;
}
}
}
else if (refo || refn || key != old_type.key)
{
if (local_semantics)
{
// If we're merging into an "unused" slot, then we
// simply accept whatever we're merging from.
if (key == unused_by_subroutine_type)
{
*this = old_type;
changed = true;
}
else if (old_type.key == unused_by_subroutine_type)
{
// Do nothing.
}
// If we already have an `unsuitable' type, then we
// don't need to change again.
else if (key != unsuitable_type)
{
key = unsuitable_type;
changed = true;
}
}
else
verifier->verify_fail ("unmergeable type");
}
return changed;
}
#ifdef VERIFY_DEBUG
void print (void) const
{
char c = '?';
switch (key)
{
case boolean_type: c = 'Z'; break;
case byte_type: c = 'B'; break;
case char_type: c = 'C'; break;
case short_type: c = 'S'; break;
case int_type: c = 'I'; break;
case long_type: c = 'J'; break;
case float_type: c = 'F'; break;
case double_type: c = 'D'; break;
case void_type: c = 'V'; break;
case unsuitable_type: c = '-'; break;
case return_address_type: c = 'r'; break;
case continuation_type: c = '+'; break;
case unused_by_subroutine_type: c = '_'; break;
case reference_type: c = 'L'; break;
case null_type: c = '@'; break;
case uninitialized_reference_type: c = 'U'; break;
}
debug_print ("%c", c);
}
#endif /* VERIFY_DEBUG */
};
// This class holds all the state information we need for a given
// location.
struct state
{
// The current top of the stack, in terms of slots.
int stacktop;
// The current depth of the stack. This will be larger than
// STACKTOP when wide types are on the stack.
int stackdepth;
// The stack.
type *stack;
// The local variables.
type *locals;
// This is used in subroutines to keep track of which local
// variables have been accessed.
bool *local_changed;
// If not 0, then we are in a subroutine. The value is the PC of
// the subroutine's entry point. We can use 0 as an exceptional
// value because PC=0 can never be a subroutine.
int subroutine;
// This is used to keep a linked list of all the states which
// require re-verification. We use the PC to keep track.
int next;
// We keep track of the type of `this' specially. This is used to
// ensure that an instance initializer invokes another initializer
// on `this' before returning. We must keep track of this
// specially because otherwise we might be confused by code which
// assigns to locals[0] (overwriting `this') and then returns
// without really initializing.
type this_type;
// This is a list of all subroutines that have been seen at this
// point. Ordinarily this is NULL; it is only allocated and used
// in relatively weird situations involving non-ret exit from a
// subroutine. We have to keep track of this in this way to avoid
// endless recursion in these cases.
subr_info *seen_subrs;
// INVALID marks a state which is not on the linked list of states
// requiring reverification.
static const int INVALID = -1;
// NO_NEXT marks the state at the end of the reverification list.
static const int NO_NEXT = -2;
// This is used to mark the stack depth at the instruction just
// after a `jsr' when we haven't yet processed the corresponding
// `ret'. See handle_jsr_insn for more information.
static const int NO_STACK = -1;
state ()
: this_type ()
{
stack = NULL;
locals = NULL;
local_changed = NULL;
seen_subrs = NULL;
}
state (int max_stack, int max_locals)
: this_type ()
{
stacktop = 0;
stackdepth = 0;
stack = new type[max_stack];
for (int i = 0; i < max_stack; ++i)
stack[i] = unsuitable_type;
locals = new type[max_locals];
local_changed = (bool *) _Jv_Malloc (sizeof (bool) * max_locals);
seen_subrs = NULL;
for (int i = 0; i < max_locals; ++i)
{
locals[i] = unsuitable_type;
local_changed[i] = false;
}
next = INVALID;
subroutine = 0;
}
state (const state *orig, int max_stack, int max_locals,
bool ret_semantics = false)
{
stack = new type[max_stack];
locals = new type[max_locals];
local_changed = (bool *) _Jv_Malloc (sizeof (bool) * max_locals);
seen_subrs = NULL;
copy (orig, max_stack, max_locals, ret_semantics);
next = INVALID;
}
~state ()
{
if (stack)
delete[] stack;
if (locals)
delete[] locals;
if (local_changed)
_Jv_Free (local_changed);
clean_subrs ();
}
void *operator new[] (size_t bytes)
{
return _Jv_Malloc (bytes);
}
void operator delete[] (void *mem)
{
_Jv_Free (mem);
}
void *operator new (size_t bytes)
{
return _Jv_Malloc (bytes);
}
void operator delete (void *mem)
{
_Jv_Free (mem);
}
void clean_subrs ()
{
subr_info *info = seen_subrs;
while (info != NULL)
{
subr_info *next = info->next;
_Jv_Free (info);
info = next;
}
}
void copy (const state *copy, int max_stack, int max_locals,
bool ret_semantics = false)
{
stacktop = copy->stacktop;
stackdepth = copy->stackdepth;
subroutine = copy->subroutine;
for (int i = 0; i < max_stack; ++i)
stack[i] = copy->stack[i];
for (int i = 0; i < max_locals; ++i)
{
// See push_jump_merge to understand this case.
if (ret_semantics)
locals[i] = type (copy->local_changed[i]
? unsuitable_type
: unused_by_subroutine_type);
else
locals[i] = copy->locals[i];
local_changed[i] = copy->local_changed[i];
}
clean_subrs ();
if (copy->seen_subrs)
{
for (subr_info *info = seen_subrs; info != NULL; info = info->next)
add_subr (info->pc);
}
else
seen_subrs = NULL;
this_type = copy->this_type;
// Don't modify `next'.
}
// Modify this state to reflect entry to an exception handler.
void set_exception (type t, int max_stack)
{
stackdepth = 1;
stacktop = 1;
stack[0] = t;
for (int i = stacktop; i < max_stack; ++i)
stack[i] = unsuitable_type;
}
// Modify this state to reflect entry into a subroutine.
void enter_subroutine (int npc, int max_locals)
{
subroutine = npc;
// Mark all items as unchanged. Each subroutine needs to keep
// track of its `changed' state independently. In the case of
// nested subroutines, this information will be merged back into
// parent by the `ret'.
for (int i = 0; i < max_locals; ++i)
local_changed[i] = false;
}
// Indicate that we've been in this this subroutine.
void add_subr (int pc)
{
subr_info *n = (subr_info *) _Jv_Malloc (sizeof (subr_info));
n->pc = pc;
n->next = seen_subrs;
seen_subrs = n;
}
// Merge STATE_OLD into this state. Destructively modifies this
// state. Returns true if the new state was in fact changed.
// Will throw an exception if the states are not mergeable.
bool merge (state *state_old, bool ret_semantics,
int max_locals, _Jv_BytecodeVerifier *verifier)
{
bool changed = false;
// Special handling for `this'. If one or the other is
// uninitialized, then the merge is uninitialized.
if (this_type.isinitialized ())
this_type = state_old->this_type;
// Merge subroutine states. Here we just keep track of what
// subroutine we think we're in. We only check for a merge
// (which is invalid) when we see a `ret'.
if (subroutine == state_old->subroutine)
{
// Nothing.
}
else if (subroutine == 0)
{
subroutine = state_old->subroutine;
changed = true;
}
else
{
// If the subroutines differ, and we haven't seen this
// subroutine before, indicate that the state changed. This
// is needed to detect when subroutines have merged.
bool found = false;
for (subr_info *info = seen_subrs; info != NULL; info = info->next)
{
if (info->pc == state_old->subroutine)
{
found = true;
break;
}
}
if (! found)
{
add_subr (state_old->subroutine);
changed = true;
}
}
// Merge stacks. Special handling for NO_STACK case.
if (state_old->stacktop == NO_STACK)
{
// Nothing to do in this case; we don't care about modifying
// the old state.
}
else if (stacktop == NO_STACK)
{
stacktop = state_old->stacktop;
stackdepth = state_old->stackdepth;
for (int i = 0; i < stacktop; ++i)
stack[i] = state_old->stack[i];
changed = true;
}
else if (state_old->stacktop != stacktop)
verifier->verify_fail ("stack sizes differ");
else
{
for (int i = 0; i < state_old->stacktop; ++i)
{
if (stack[i].merge (state_old->stack[i], false, verifier))
changed = true;
}
}
// Merge local variables.
for (int i = 0; i < max_locals; ++i)
{
// If we're not processing a `ret', then we merge every
// local variable. If we are processing a `ret', then we
// only merge locals which changed in the subroutine. When
// processing a `ret', STATE_OLD is the state at the point
// of the `ret', and THIS is the state just after the `jsr'.
if (! ret_semantics || state_old->local_changed[i])
{
if (locals[i].merge (state_old->locals[i], true, verifier))
{
// Note that we don't call `note_variable' here.
// This change doesn't represent a real change to a
// local, but rather a merge artifact. If we're in
// a subroutine which is called with two
// incompatible types in a slot that is unused by
// the subroutine, then we don't want to mark that
// variable as having been modified.
changed = true;
}
}
// If we're in a subroutine, we must compute the union of
// all the changed local variables.
if (state_old->local_changed[i])
note_variable (i);
}
return changed;
}
// Throw an exception if there is an uninitialized object on the
// stack or in a local variable. EXCEPTION_SEMANTICS controls
// whether we're using backwards-branch or exception-handing
// semantics.
void check_no_uninitialized_objects (int max_locals,
_Jv_BytecodeVerifier *verifier,
bool exception_semantics = false)
{
if (! exception_semantics)
{
for (int i = 0; i < stacktop; ++i)
if (stack[i].isreference () && ! stack[i].isinitialized ())
verifier->verify_fail ("uninitialized object on stack");
}
for (int i = 0; i < max_locals; ++i)
if (locals[i].isreference () && ! locals[i].isinitialized ())
verifier->verify_fail ("uninitialized object in local variable");
check_this_initialized (verifier);
}
// Ensure that `this' has been initialized.
void check_this_initialized (_Jv_BytecodeVerifier *verifier)
{
if (this_type.isreference () && ! this_type.isinitialized ())
verifier->verify_fail ("`this' is uninitialized");
}
// Set type of `this'.
void set_this_type (const type &k)
{
this_type = k;
}
// Note that a local variable was modified.
void note_variable (int index)
{
if (subroutine > 0)
local_changed[index] = true;
}
// Mark each `new'd object we know of that was allocated at PC as
// initialized.
void set_initialized (int pc, int max_locals)
{
for (int i = 0; i < stacktop; ++i)
stack[i].set_initialized (pc);
for (int i = 0; i < max_locals; ++i)
locals[i].set_initialized (pc);
this_type.set_initialized (pc);
}
// Return true if this state is the unmerged result of a `ret'.
bool is_unmerged_ret_state (int max_locals) const
{
if (stacktop == NO_STACK)
return true;
for (int i = 0; i < max_locals; ++i)
{
if (locals[i].key == unused_by_subroutine_type)
return true;
}
return false;
}
#ifdef VERIFY_DEBUG
void print (const char *leader, int pc,
int max_stack, int max_locals) const
{
debug_print ("%s [%4d]: [stack] ", leader, pc);
int i;
for (i = 0; i < stacktop; ++i)
stack[i].print ();
for (; i < max_stack; ++i)
debug_print (".");
debug_print (" [local] ");
for (i = 0; i < max_locals; ++i)
{
locals[i].print ();
debug_print (local_changed[i] ? "+" : " ");
}
if (subroutine == 0)
debug_print (" | None");
else
debug_print (" | %4d", subroutine);
debug_print (" | %p\n", this);
}
#else
inline void print (const char *, int, int, int) const
{
}
#endif /* VERIFY_DEBUG */
};
type pop_raw ()
{
if (current_state->stacktop <= 0)
verify_fail ("stack empty");
type r = current_state->stack[--current_state->stacktop];
current_state->stackdepth -= r.depth ();
if (current_state->stackdepth < 0)
verify_fail ("stack empty", start_PC);
return r;
}
type pop32 ()
{
type r = pop_raw ();
if (r.iswide ())
verify_fail ("narrow pop of wide type");
return r;
}
type pop_type (type match)
{
match.promote ();
type t = pop_raw ();
if (! match.compatible (t, this))
verify_fail ("incompatible type on stack");
return t;
}
// Pop a reference which is guaranteed to be initialized. MATCH
// doesn't have to be a reference type; in this case this acts like
// pop_type.
type pop_init_ref (type match)
{
type t = pop_raw ();
if (t.isreference () && ! t.isinitialized ())
verify_fail ("initialized reference required");
else if (! match.compatible (t, this))
verify_fail ("incompatible type on stack");
return t;
}
// Pop a reference type or a return address.
type pop_ref_or_return ()
{
type t = pop_raw ();
if (! t.isreference () && t.key != return_address_type)
verify_fail ("expected reference or return address on stack");
return t;
}
void push_type (type t)
{
// If T is a numeric type like short, promote it to int.
t.promote ();
int depth = t.depth ();
if (current_state->stackdepth + depth > current_method->max_stack)
verify_fail ("stack overflow");
current_state->stack[current_state->stacktop++] = t;
current_state->stackdepth += depth;
}
void set_variable (int index, type t)
{
// If T is a numeric type like short, promote it to int.
t.promote ();
int depth = t.depth ();
if (index > current_method->max_locals - depth)
verify_fail ("invalid local variable");
current_state->locals[index] = t;
current_state->note_variable (index);
if (depth == 2)
{
current_state->locals[index + 1] = continuation_type;
current_state->note_variable (index + 1);
}
if (index > 0 && current_state->locals[index - 1].iswide ())
{
current_state->locals[index - 1] = unsuitable_type;
// There's no need to call note_variable here.
}
}
type get_variable (int index, type t)
{
int depth = t.depth ();
if (index > current_method->max_locals - depth)
verify_fail ("invalid local variable");
if (! t.compatible (current_state->locals[index], this))
verify_fail ("incompatible type in local variable");
if (depth == 2)
{
type t (continuation_type);
if (! current_state->locals[index + 1].compatible (t, this))
verify_fail ("invalid local variable");
}
return current_state->locals[index];
}
// Make sure ARRAY is an array type and that its elements are
// compatible with type ELEMENT. Returns the actual element type.
type require_array_type (type array, type element)
{
// An odd case. Here we just pretend that everything went ok. If
// the requested element type is some kind of reference, return
// the null type instead.
if (array.isnull ())
return element.isreference () ? type (null_type) : element;
if (! array.isarray ())
verify_fail ("array required");
type t = array.element_type (this);
if (! element.compatible (t, this))
{
// Special case for byte arrays, which must also be boolean
// arrays.
bool ok = true;
if (element.key == byte_type)
{
type e2 (boolean_type);
ok = e2.compatible (t, this);
}
if (! ok)
verify_fail ("incompatible array element type");
}
// Return T and not ELEMENT, because T might be specialized.
return t;
}
jint get_byte ()
{
if (PC >= current_method->code_length)
verify_fail ("premature end of bytecode");
return (jint) bytecode[PC++] & 0xff;
}
jint get_ushort ()
{
jint b1 = get_byte ();
jint b2 = get_byte ();
return (jint) ((b1 << 8) | b2) & 0xffff;
}
jint get_short ()
{
jint b1 = get_byte ();
jint b2 = get_byte ();
jshort s = (b1 << 8) | b2;
return (jint) s;
}
jint get_int ()
{
jint b1 = get_byte ();
jint b2 = get_byte ();
jint b3 = get_byte ();
jint b4 = get_byte ();
return (b1 << 24) | (b2 << 16) | (b3 << 8) | b4;
}
int compute_jump (int offset)
{
int npc = start_PC + offset;
if (npc < 0 || npc >= current_method->code_length)
verify_fail ("branch out of range", start_PC);
return npc;
}
// Merge the indicated state into the state at the branch target and
// schedule a new PC if there is a change. If RET_SEMANTICS is
// true, then we are merging from a `ret' instruction into the
// instruction after a `jsr'. This is a special case with its own
// modified semantics.
void push_jump_merge (int npc, state *nstate, bool ret_semantics = false)
{
bool changed = true;
if (states[npc] == NULL)
{
// There's a weird situation here. If are examining the
// branch that results from a `ret', and there is not yet a
// state available at the branch target (the instruction just
// after the `jsr'), then we have to construct a special kind
// of state at that point for future merging. This special
// state has the type `unused_by_subroutine_type' in each slot
// which was not modified by the subroutine.
states[npc] = new state (nstate, current_method->max_stack,
current_method->max_locals, ret_semantics);
debug_print ("== New state in push_jump_merge\n");
states[npc]->print ("New", npc, current_method->max_stack,
current_method->max_locals);
}
else
{
debug_print ("== Merge states in push_jump_merge\n");
nstate->print ("Frm", start_PC, current_method->max_stack,
current_method->max_locals);
states[npc]->print (" To", npc, current_method->max_stack,
current_method->max_locals);
changed = states[npc]->merge (nstate, ret_semantics,
current_method->max_locals, this);
states[npc]->print ("New", npc, current_method->max_stack,
current_method->max_locals);
}
if (changed && states[npc]->next == state::INVALID)
{
// The merge changed the state, and the new PC isn't yet on our
// list of PCs to re-verify.
states[npc]->next = next_verify_pc;
next_verify_pc = npc;
}
}
void push_jump (int offset)
{
int npc = compute_jump (offset);
if (npc < PC)
current_state->check_no_uninitialized_objects (current_method->max_locals, this);
push_jump_merge (npc, current_state);
}
void push_exception_jump (type t, int pc)
{
current_state->check_no_uninitialized_objects (current_method->max_locals,
this, true);
state s (current_state, current_method->max_stack,
current_method->max_locals);
if (current_method->max_stack < 1)
verify_fail ("stack overflow at exception handler");
s.set_exception (t, current_method->max_stack);
push_jump_merge (pc, &s);
}
int pop_jump ()
{
int *prev_loc = &next_verify_pc;
int npc = next_verify_pc;
while (npc != state::NO_NEXT)
{
// If the next available PC is an unmerged `ret' state, then
// we aren't yet ready to handle it. That's because we would
// need all kind of special cases to do so. So instead we
// defer this jump until after we've processed it via a
// fall-through. This has to happen because the instruction
// before this one must be a `jsr'.
if (! states[npc]->is_unmerged_ret_state (current_method->max_locals))
{
*prev_loc = states[npc]->next;
states[npc]->next = state::INVALID;
return npc;
}
prev_loc = &states[npc]->next;
npc = states[npc]->next;
}
// Note that we might have gotten here even when there are
// remaining states to process. That can happen if we find a
// `jsr' without a `ret'.
return state::NO_NEXT;
}
void invalidate_pc ()
{
PC = state::NO_NEXT;
}
void note_branch_target (int pc, bool is_jsr_target = false)
{
// Don't check `pc <= PC', because we've advanced PC after
// fetching the target and we haven't yet checked the next
// instruction.
if (pc < PC && ! (flags[pc] & FLAG_INSN_START))
verify_fail ("branch not to instruction start", start_PC);
flags[pc] |= FLAG_BRANCH_TARGET;
if (is_jsr_target)
{
// Record the jsr which called this instruction.
subr_info *info = (subr_info *) _Jv_Malloc (sizeof (subr_info));
info->pc = PC;
info->next = jsr_ptrs[pc];
jsr_ptrs[pc] = info;
}
}
void skip_padding ()
{
while ((PC % 4) > 0)
if (get_byte () != 0)
verify_fail ("found nonzero padding byte");
}
// Return the subroutine to which the instruction at PC belongs.
int get_subroutine (int pc)
{
if (states[pc] == NULL)
return 0;
return states[pc]->subroutine;
}
// Do the work for a `ret' instruction. INDEX is the index into the
// local variables.
void handle_ret_insn (int index)
{
get_variable (index, return_address_type);
int csub = current_state->subroutine;
if (csub == 0)
verify_fail ("no subroutine");
// Check to see if we've merged subroutines.
subr_entry_info *entry;
for (entry = entry_points; entry != NULL; entry = entry->next)
{
if (entry->ret_pc == start_PC)
break;
}
if (entry == NULL)
{
entry = (subr_entry_info *) _Jv_Malloc (sizeof (subr_entry_info));
entry->pc = csub;
entry->ret_pc = start_PC;
entry->next = entry_points;
entry_points = entry;
}
else if (entry->pc != csub)
verify_fail ("subroutines merged");
for (subr_info *subr = jsr_ptrs[csub]; subr != NULL; subr = subr->next)
{
// We might be returning to a `jsr' that is at the end of the
// bytecode. This is ok if we never return from the called
// subroutine, but if we see this here it is an error.
if (subr->pc >= current_method->code_length)
verify_fail ("fell off end");
// Temporarily modify the current state so it looks like we're
// in the enclosing context.
current_state->subroutine = get_subroutine (subr->pc);
if (subr->pc < PC)
current_state->check_no_uninitialized_objects (current_method->max_locals, this);
push_jump_merge (subr->pc, current_state, true);
}
current_state->subroutine = csub;
invalidate_pc ();
}
// We're in the subroutine SUB, calling a subroutine at DEST. Make
// sure this subroutine isn't already on the stack.
void check_nonrecursive_call (int sub, int dest)
{
if (sub == 0)
return;
if (sub == dest)
verify_fail ("recursive subroutine call");
for (subr_info *info = jsr_ptrs[sub]; info != NULL; info = info->next)
check_nonrecursive_call (get_subroutine (info->pc), dest);
}
void handle_jsr_insn (int offset)
{
int npc = compute_jump (offset);
if (npc < PC)
current_state->check_no_uninitialized_objects (current_method->max_locals, this);
check_nonrecursive_call (current_state->subroutine, npc);
// Modify our state as appropriate for entry into a subroutine.
push_type (return_address_type);
push_jump_merge (npc, current_state);
// Clean up.
pop_type (return_address_type);
// On entry to the subroutine, the subroutine number must be set
// and the locals must be marked as cleared. We do this after
// merging state so that we don't erroneously "notice" a variable
// change merely on entry.
states[npc]->enter_subroutine (npc, current_method->max_locals);
// Indicate that we don't know the stack depth of the instruction
// following the `jsr'. The idea here is that we need to merge
// the local variable state across the jsr, but the subroutine
// might change the stack depth, so we can't make any assumptions
// about it. So we have yet another special case. We know that
// at this point PC points to the instruction after the jsr. Note
// that it is ok to have a `jsr' at the end of the bytecode,
// provided that the called subroutine never returns. So, we have
// a special case here and another one when we handle the ret.
if (PC < current_method->code_length)
{
current_state->stacktop = state::NO_STACK;
push_jump_merge (PC, current_state);
}
invalidate_pc ();
}
jclass construct_primitive_array_type (type_val prim)
{
jclass k = NULL;
switch (prim)
{
case boolean_type:
k = JvPrimClass (boolean);
break;
case char_type:
k = JvPrimClass (char);
break;
case float_type:
k = JvPrimClass (float);
break;
case double_type:
k = JvPrimClass (double);
break;
case byte_type:
k = JvPrimClass (byte);
break;
case short_type:
k = JvPrimClass (short);
break;
case int_type:
k = JvPrimClass (int);
break;
case long_type:
k = JvPrimClass (long);
break;
// These aren't used here but we call them out to avoid
// warnings.
case void_type:
case unsuitable_type:
case return_address_type:
case continuation_type:
case unused_by_subroutine_type:
case reference_type:
case null_type:
case uninitialized_reference_type:
default:
verify_fail ("unknown type in construct_primitive_array_type");
}
k = _Jv_GetArrayClass (k, NULL);
return k;
}
// This pass computes the location of branch targets and also
// instruction starts.
void branch_prepass ()
{
flags = (char *) _Jv_Malloc (current_method->code_length);
jsr_ptrs = (subr_info **) _Jv_Malloc (sizeof (subr_info *)
* current_method->code_length);
for (int i = 0; i < current_method->code_length; ++i)
{
flags[i] = 0;
jsr_ptrs[i] = NULL;
}
bool last_was_jsr = false;
PC = 0;
while (PC < current_method->code_length)
{
// Set `start_PC' early so that error checking can have the
// correct value.
start_PC = PC;
flags[PC] |= FLAG_INSN_START;
// If the previous instruction was a jsr, then the next
// instruction is a branch target -- the branch being the
// corresponding `ret'.
if (last_was_jsr)
note_branch_target (PC);
last_was_jsr = false;
java_opcode opcode = (java_opcode) bytecode[PC++];
switch (opcode)
{
case op_nop:
case op_aconst_null:
case op_iconst_m1:
case op_iconst_0:
case op_iconst_1:
case op_iconst_2:
case op_iconst_3:
case op_iconst_4:
case op_iconst_5:
case op_lconst_0:
case op_lconst_1:
case op_fconst_0:
case op_fconst_1:
case op_fconst_2:
case op_dconst_0:
case op_dconst_1:
case op_iload_0:
case op_iload_1:
case op_iload_2:
case op_iload_3:
case op_lload_0:
case op_lload_1:
case op_lload_2:
case op_lload_3:
case op_fload_0:
case op_fload_1:
case op_fload_2:
case op_fload_3:
case op_dload_0:
case op_dload_1:
case op_dload_2:
case op_dload_3:
case op_aload_0:
case op_aload_1:
case op_aload_2:
case op_aload_3:
case op_iaload:
case op_laload:
case op_faload:
case op_daload:
case op_aaload:
case op_baload:
case op_caload:
case op_saload:
case op_istore_0:
case op_istore_1:
case op_istore_2:
case op_istore_3:
case op_lstore_0:
case op_lstore_1:
case op_lstore_2:
case op_lstore_3:
case op_fstore_0:
case op_fstore_1:
case op_fstore_2:
case op_fstore_3:
case op_dstore_0:
case op_dstore_1:
case op_dstore_2:
case op_dstore_3:
case op_astore_0:
case op_astore_1:
case op_astore_2:
case op_astore_3:
case op_iastore:
case op_lastore:
case op_fastore:
case op_dastore:
case op_aastore:
case op_bastore:
case op_castore:
case op_sastore:
case op_pop:
case op_pop2:
case op_dup:
case op_dup_x1:
case op_dup_x2:
case op_dup2:
case op_dup2_x1:
case op_dup2_x2:
case op_swap:
case op_iadd:
case op_isub:
case op_imul:
case op_idiv:
case op_irem:
case op_ishl:
case op_ishr:
case op_iushr:
case op_iand:
case op_ior:
case op_ixor:
case op_ladd:
case op_lsub:
case op_lmul:
case op_ldiv:
case op_lrem:
case op_lshl:
case op_lshr:
case op_lushr:
case op_land:
case op_lor:
case op_lxor:
case op_fadd:
case op_fsub:
case op_fmul:
case op_fdiv:
case op_frem:
case op_dadd:
case op_dsub:
case op_dmul:
case op_ddiv:
case op_drem:
case op_ineg:
case op_i2b:
case op_i2c:
case op_i2s:
case op_lneg:
case op_fneg:
case op_dneg:
case op_i2l:
case op_i2f:
case op_i2d:
case op_l2i:
case op_l2f:
case op_l2d:
case op_f2i:
case op_f2l:
case op_f2d:
case op_d2i:
case op_d2l:
case op_d2f:
case op_lcmp:
case op_fcmpl:
case op_fcmpg:
case op_dcmpl:
case op_dcmpg:
case op_monitorenter:
case op_monitorexit:
case op_ireturn:
case op_lreturn:
case op_freturn:
case op_dreturn:
case op_areturn:
case op_return:
case op_athrow:
case op_arraylength:
break;
case op_bipush:
case op_ldc:
case op_iload:
case op_lload:
case op_fload:
case op_dload:
case op_aload:
case op_istore:
case op_lstore:
case op_fstore:
case op_dstore:
case op_astore:
case op_ret:
case op_newarray:
get_byte ();
break;
case op_iinc:
case op_sipush:
case op_ldc_w:
case op_ldc2_w:
case op_getstatic:
case op_getfield:
case op_putfield:
case op_putstatic:
case op_new:
case op_anewarray:
case op_instanceof:
case op_checkcast:
case op_invokespecial:
case op_invokestatic:
case op_invokevirtual:
get_short ();
break;
case op_multianewarray:
get_short ();
get_byte ();
break;
case op_jsr:
last_was_jsr = true;
// Fall through.
case op_ifeq:
case op_ifne:
case op_iflt:
case op_ifge:
case op_ifgt:
case op_ifle:
case op_if_icmpeq:
case op_if_icmpne:
case op_if_icmplt:
case op_if_icmpge:
case op_if_icmpgt:
case op_if_icmple:
case op_if_acmpeq:
case op_if_acmpne:
case op_ifnull:
case op_ifnonnull:
case op_goto:
note_branch_target (compute_jump (get_short ()), last_was_jsr);
break;
case op_tableswitch:
{
skip_padding ();
note_branch_target (compute_jump (get_int ()));
jint low = get_int ();
jint hi = get_int ();
if (low > hi)
verify_fail ("invalid tableswitch", start_PC);
for (int i = low; i <= hi; ++i)
note_branch_target (compute_jump (get_int ()));
}
break;
case op_lookupswitch:
{
skip_padding ();
note_branch_target (compute_jump (get_int ()));
int npairs = get_int ();
if (npairs < 0)
verify_fail ("too few pairs in lookupswitch", start_PC);
while (npairs-- > 0)
{
get_int ();
note_branch_target (compute_jump (get_int ()));
}
}
break;
case op_invokeinterface:
get_short ();
get_byte ();
get_byte ();
break;
case op_wide:
{
opcode = (java_opcode) get_byte ();
get_short ();
if (opcode == op_iinc)
get_short ();
}
break;
case op_jsr_w:
last_was_jsr = true;
// Fall through.
case op_goto_w:
note_branch_target (compute_jump (get_int ()), last_was_jsr);
break;
// These are unused here, but we call them out explicitly
// so that -Wswitch-enum doesn't complain.
case op_putfield_1:
case op_putfield_2:
case op_putfield_4:
case op_putfield_8:
case op_putfield_a:
case op_putstatic_1:
case op_putstatic_2:
case op_putstatic_4:
case op_putstatic_8:
case op_putstatic_a:
case op_getfield_1:
case op_getfield_2s:
case op_getfield_2u:
case op_getfield_4:
case op_getfield_8:
case op_getfield_a:
case op_getstatic_1:
case op_getstatic_2s:
case op_getstatic_2u:
case op_getstatic_4:
case op_getstatic_8:
case op_getstatic_a:
default:
verify_fail ("unrecognized instruction in branch_prepass",
start_PC);
}
// See if any previous branch tried to branch to the middle of
// this instruction.
for (int pc = start_PC + 1; pc < PC; ++pc)
{
if ((flags[pc] & FLAG_BRANCH_TARGET))
verify_fail ("branch to middle of instruction", pc);
}
}
// Verify exception handlers.
for (int i = 0; i < current_method->exc_count; ++i)
{
if (! (flags[exception[i].handler_pc.i] & FLAG_INSN_START))
verify_fail ("exception handler not at instruction start",
exception[i].handler_pc.i);
if (! (flags[exception[i].start_pc.i] & FLAG_INSN_START))
verify_fail ("exception start not at instruction start",
exception[i].start_pc.i);
if (exception[i].end_pc.i != current_method->code_length
&& ! (flags[exception[i].end_pc.i] & FLAG_INSN_START))
verify_fail ("exception end not at instruction start",
exception[i].end_pc.i);
flags[exception[i].handler_pc.i] |= FLAG_BRANCH_TARGET;
}
}
void check_pool_index (int index)
{
if (index < 0 || index >= current_class->constants.size)
verify_fail ("constant pool index out of range", start_PC);
}
type check_class_constant (int index)
{
check_pool_index (index);
_Jv_Constants *pool = &current_class->constants;
if (pool->tags[index] == JV_CONSTANT_ResolvedClass)
return type (pool->data[index].clazz, this);
else if (pool->tags[index] == JV_CONSTANT_Class)
return type (pool->data[index].utf8, this);
verify_fail ("expected class constant", start_PC);
}
type check_constant (int index)
{
check_pool_index (index);
_Jv_Constants *pool = &current_class->constants;
if (pool->tags[index] == JV_CONSTANT_ResolvedString
|| pool->tags[index] == JV_CONSTANT_String)
return type (&java::lang::String::class$, this);
else if (pool->tags[index] == JV_CONSTANT_Integer)
return type (int_type);
else if (pool->tags[index] == JV_CONSTANT_Float)
return type (float_type);
verify_fail ("String, int, or float constant expected", start_PC);
}
type check_wide_constant (int index)
{
check_pool_index (index);
_Jv_Constants *pool = &current_class->constants;
if (pool->tags[index] == JV_CONSTANT_Long)
return type (long_type);
else if (pool->tags[index] == JV_CONSTANT_Double)
return type (double_type);
verify_fail ("long or double constant expected", start_PC);
}
// Helper for both field and method. These are laid out the same in
// the constant pool.
type handle_field_or_method (int index, int expected,
_Jv_Utf8Const **name,
_Jv_Utf8Const **fmtype)
{
check_pool_index (index);
_Jv_Constants *pool = &current_class->constants;
if (pool->tags[index] != expected)
verify_fail ("didn't see expected constant", start_PC);
// Once we know we have a Fieldref or Methodref we assume that it
// is correctly laid out in the constant pool. I think the code
// in defineclass.cc guarantees this.
_Jv_ushort class_index, name_and_type_index;
_Jv_loadIndexes (&pool->data[index],
class_index,
name_and_type_index);
_Jv_ushort name_index, desc_index;
_Jv_loadIndexes (&pool->data[name_and_type_index],
name_index, desc_index);
*name = pool->data[name_index].utf8;
*fmtype = pool->data[desc_index].utf8;
return check_class_constant (class_index);
}
// Return field's type, compute class' type if requested.
type check_field_constant (int index, type *class_type = NULL)
{
_Jv_Utf8Const *name, *field_type;
type ct = handle_field_or_method (index,
JV_CONSTANT_Fieldref,
&name, &field_type);
if (class_type)
*class_type = ct;
if (field_type->data[0] == '[' || field_type->data[0] == 'L')
return type (field_type, this);
return get_type_val_for_signature (field_type->data[0]);
}
type check_method_constant (int index, bool is_interface,
_Jv_Utf8Const **method_name,
_Jv_Utf8Const **method_signature)
{
return handle_field_or_method (index,
(is_interface
? JV_CONSTANT_InterfaceMethodref
: JV_CONSTANT_Methodref),
method_name, method_signature);
}
type get_one_type (char *&p)
{
char *start = p;
int arraycount = 0;
while (*p == '[')
{
++arraycount;
++p;
}
char v = *p++;
if (v == 'L')
{
while (*p != ';')
++p;
++p;
_Jv_Utf8Const *name = make_utf8_const (start, p - start);
return type (name, this);
}
// Casting to jchar here is ok since we are looking at an ASCII
// character.
type_val rt = get_type_val_for_signature (jchar (v));
if (arraycount == 0)
{
// Callers of this function eventually push their arguments on
// the stack. So, promote them here.
return type (rt).promote ();
}
jclass k = construct_primitive_array_type (rt);
while (--arraycount > 0)
k = _Jv_GetArrayClass (k, NULL);
return type (k, this);
}
void compute_argument_types (_Jv_Utf8Const *signature,
type *types)
{
char *p = signature->data;
// Skip `('.
++p;
int i = 0;
while (*p != ')')
types[i++] = get_one_type (p);
}
type compute_return_type (_Jv_Utf8Const *signature)
{
char *p = signature->data;
while (*p != ')')
++p;
++p;
return get_one_type (p);
}
void check_return_type (type onstack)
{
type rt = compute_return_type (current_method->self->signature);
if (! rt.compatible (onstack, this))
verify_fail ("incompatible return type");
}
// Initialize the stack for the new method. Returns true if this
// method is an instance initializer.
bool initialize_stack ()
{
int var = 0;
bool is_init = _Jv_equalUtf8Consts (current_method->self->name,
gcj::init_name);
bool is_clinit = _Jv_equalUtf8Consts (current_method->self->name,
gcj::clinit_name);
using namespace java::lang::reflect;
if (! Modifier::isStatic (current_method->self->accflags))
{
type kurr (current_class, this);
if (is_init)
{
kurr.set_uninitialized (type::SELF, this);
is_init = true;
}
else if (is_clinit)
verify_fail ("<clinit> method must be static");
set_variable (0, kurr);
current_state->set_this_type (kurr);
++var;
}
else
{
if (is_init)
verify_fail ("<init> method must be non-static");
}
// We have to handle wide arguments specially here.
int arg_count = _Jv_count_arguments (current_method->self->signature);
type arg_types[arg_count];
compute_argument_types (current_method->self->signature, arg_types);
for (int i = 0; i < arg_count; ++i)
{
set_variable (var, arg_types[i]);
++var;
if (arg_types[i].iswide ())
++var;
}
return is_init;
}
void verify_instructions_0 ()
{
current_state = new state (current_method->max_stack,
current_method->max_locals);
PC = 0;
start_PC = 0;
// True if we are verifying an instance initializer.
bool this_is_init = initialize_stack ();
states = (state **) _Jv_Malloc (sizeof (state *)
* current_method->code_length);
for (int i = 0; i < current_method->code_length; ++i)
states[i] = NULL;
next_verify_pc = state::NO_NEXT;
while (true)
{
// If the PC was invalidated, get a new one from the work list.
if (PC == state::NO_NEXT)
{
PC = pop_jump ();
if (PC == state::INVALID)
verify_fail ("can't happen: saw state::INVALID");
if (PC == state::NO_NEXT)
break;
debug_print ("== State pop from pending list\n");
// Set up the current state.
current_state->copy (states[PC], current_method->max_stack,
current_method->max_locals);
}
else
{
// Control can't fall off the end of the bytecode. We
// only need to check this in the fall-through case,
// because branch bounds are checked when they are
// pushed.
if (PC >= current_method->code_length)
verify_fail ("fell off end");
// We only have to do this checking in the situation where
// control flow falls through from the previous
// instruction. Otherwise merging is done at the time we
// push the branch.
if (states[PC] != NULL)
{
// We've already visited this instruction. So merge
// the states together. If this yields no change then
// we don't have to re-verify. However, if the new
// state is an the result of an unmerged `ret', we
// must continue through it.
debug_print ("== Fall through merge\n");
states[PC]->print ("Old", PC, current_method->max_stack,
current_method->max_locals);
current_state->print ("Cur", PC, current_method->max_stack,
current_method->max_locals);
if (! current_state->merge (states[PC], false,
current_method->max_locals, this)
&& ! states[PC]->is_unmerged_ret_state (current_method->max_locals))
{
debug_print ("== Fall through optimization\n");
invalidate_pc ();
continue;
}
// Save a copy of it for later.
states[PC]->copy (current_state, current_method->max_stack,
current_method->max_locals);
current_state->print ("New", PC, current_method->max_stack,
current_method->max_locals);
}
}
// We only have to keep saved state at branch targets. If
// we're at a branch target and the state here hasn't been set
// yet, we set it now.
if (states[PC] == NULL && (flags[PC] & FLAG_BRANCH_TARGET))
{
states[PC] = new state (current_state, current_method->max_stack,
current_method->max_locals);
}
// Set this before handling exceptions so that debug output is
// sane.
start_PC = PC;
// Update states for all active exception handlers. Ordinarily
// there are not many exception handlers. So we simply run
// through them all.
for (int i = 0; i < current_method->exc_count; ++i)
{
if (PC >= exception[i].start_pc.i && PC < exception[i].end_pc.i)
{
type handler (&java::lang::Throwable::class$, this);
if (exception[i].handler_type.i != 0)
handler = check_class_constant (exception[i].handler_type.i);
push_exception_jump (handler, exception[i].handler_pc.i);
}
}
current_state->print (" ", PC, current_method->max_stack,
current_method->max_locals);
java_opcode opcode = (java_opcode) bytecode[PC++];
switch (opcode)
{
case op_nop:
break;
case op_aconst_null:
push_type (null_type);
break;
case op_iconst_m1:
case op_iconst_0:
case op_iconst_1:
case op_iconst_2:
case op_iconst_3:
case op_iconst_4:
case op_iconst_5:
push_type (int_type);
break;
case op_lconst_0:
case op_lconst_1:
push_type (long_type);
break;
case op_fconst_0:
case op_fconst_1:
case op_fconst_2:
push_type (float_type);
break;
case op_dconst_0:
case op_dconst_1:
push_type (double_type);
break;
case op_bipush:
get_byte ();
push_type (int_type);
break;
case op_sipush:
get_short ();
push_type (int_type);
break;
case op_ldc:
push_type (check_constant (get_byte ()));
break;
case op_ldc_w:
push_type (check_constant (get_ushort ()));
break;
case op_ldc2_w:
push_type (check_wide_constant (get_ushort ()));
break;
case op_iload:
push_type (get_variable (get_byte (), int_type));
break;
case op_lload:
push_type (get_variable (get_byte (), long_type));
break;
case op_fload:
push_type (get_variable (get_byte (), float_type));
break;
case op_dload:
push_type (get_variable (get_byte (), double_type));
break;
case op_aload:
push_type (get_variable (get_byte (), reference_type));
break;
case op_iload_0:
case op_iload_1:
case op_iload_2:
case op_iload_3:
push_type (get_variable (opcode - op_iload_0, int_type));
break;
case op_lload_0:
case op_lload_1:
case op_lload_2:
case op_lload_3:
push_type (get_variable (opcode - op_lload_0, long_type));
break;
case op_fload_0:
case op_fload_1:
case op_fload_2:
case op_fload_3:
push_type (get_variable (opcode - op_fload_0, float_type));
break;
case op_dload_0:
case op_dload_1:
case op_dload_2:
case op_dload_3:
push_type (get_variable (opcode - op_dload_0, double_type));
break;
case op_aload_0:
case op_aload_1:
case op_aload_2:
case op_aload_3:
push_type (get_variable (opcode - op_aload_0, reference_type));
break;
case op_iaload:
pop_type (int_type);
push_type (require_array_type (pop_init_ref (reference_type),
int_type));
break;
case op_laload:
pop_type (int_type);
push_type (require_array_type (pop_init_ref (reference_type),
long_type));
break;
case op_faload:
pop_type (int_type);
push_type (require_array_type (pop_init_ref (reference_type),
float_type));
break;
case op_daload:
pop_type (int_type);
push_type (require_array_type (pop_init_ref (reference_type),
double_type));
break;
case op_aaload:
pop_type (int_type);
push_type (require_array_type (pop_init_ref (reference_type),
reference_type));
break;
case op_baload:
pop_type (int_type);
require_array_type (pop_init_ref (reference_type), byte_type);
push_type (int_type);
break;
case op_caload:
pop_type (int_type);
require_array_type (pop_init_ref (reference_type), char_type);
push_type (int_type);
break;
case op_saload:
pop_type (int_type);
require_array_type (pop_init_ref (reference_type), short_type);
push_type (int_type);
break;
case op_istore:
set_variable (get_byte (), pop_type (int_type));
break;
case op_lstore:
set_variable (get_byte (), pop_type (long_type));
break;
case op_fstore:
set_variable (get_byte (), pop_type (float_type));
break;
case op_dstore:
set_variable (get_byte (), pop_type (double_type));
break;
case op_astore:
set_variable (get_byte (), pop_ref_or_return ());
break;
case op_istore_0:
case op_istore_1:
case op_istore_2:
case op_istore_3:
set_variable (opcode - op_istore_0, pop_type (int_type));
break;
case op_lstore_0:
case op_lstore_1:
case op_lstore_2:
case op_lstore_3:
set_variable (opcode - op_lstore_0, pop_type (long_type));
break;
case op_fstore_0:
case op_fstore_1:
case op_fstore_2:
case op_fstore_3:
set_variable (opcode - op_fstore_0, pop_type (float_type));
break;
case op_dstore_0:
case op_dstore_1:
case op_dstore_2:
case op_dstore_3:
set_variable (opcode - op_dstore_0, pop_type (double_type));
break;
case op_astore_0:
case op_astore_1:
case op_astore_2:
case op_astore_3:
set_variable (opcode - op_astore_0, pop_ref_or_return ());
break;
case op_iastore:
pop_type (int_type);
pop_type (int_type);
require_array_type (pop_init_ref (reference_type), int_type);
break;
case op_lastore:
pop_type (long_type);
pop_type (int_type);
require_array_type (pop_init_ref (reference_type), long_type);
break;
case op_fastore:
pop_type (float_type);
pop_type (int_type);
require_array_type (pop_init_ref (reference_type), float_type);
break;
case op_dastore:
pop_type (double_type);
pop_type (int_type);
require_array_type (pop_init_ref (reference_type), double_type);
break;
case op_aastore:
pop_type (reference_type);
pop_type (int_type);
require_array_type (pop_init_ref (reference_type), reference_type);
break;
case op_bastore:
pop_type (int_type);
pop_type (int_type);
require_array_type (pop_init_ref (reference_type), byte_type);
break;
case op_castore:
pop_type (int_type);
pop_type (int_type);
require_array_type (pop_init_ref (reference_type), char_type);
break;
case op_sastore:
pop_type (int_type);
pop_type (int_type);
require_array_type (pop_init_ref (reference_type), short_type);
break;
case op_pop:
pop32 ();
break;
case op_pop2:
{
type t = pop_raw ();
if (! t.iswide ())
pop32 ();
}
break;
case op_dup:
{
type t = pop32 ();
push_type (t);
push_type (t);
}
break;
case op_dup_x1:
{
type t1 = pop32 ();
type t2 = pop32 ();
push_type (t1);
push_type (t2);
push_type (t1);
}
break;
case op_dup_x2:
{
type t1 = pop32 ();
type t2 = pop_raw ();
if (! t2.iswide ())
{
type t3 = pop32 ();
push_type (t1);
push_type (t3);
}
else
push_type (t1);
push_type (t2);
push_type (t1);
}
break;
case op_dup2:
{
type t = pop_raw ();
if (! t.iswide ())
{
type t2 = pop32 ();
push_type (t2);
push_type (t);
push_type (t2);
}
else
push_type (t);
push_type (t);
}
break;
case op_dup2_x1:
{
type t1 = pop_raw ();
type t2 = pop32 ();
if (! t1.iswide ())
{
type t3 = pop32 ();
push_type (t2);
push_type (t1);
push_type (t3);
}
else
push_type (t1);
push_type (t2);
push_type (t1);
}
break;
case op_dup2_x2:
{
type t1 = pop_raw ();
if (t1.iswide ())
{
type t2 = pop_raw ();
if (t2.iswide ())
{
push_type (t1);
push_type (t2);
}
else
{
type t3 = pop32 ();
push_type (t1);
push_type (t3);
push_type (t2);
}
push_type (t1);
}
else
{
type t2 = pop32 ();
type t3 = pop_raw ();
if (t3.iswide ())
{
push_type (t2);
push_type (t1);
}
else
{
type t4 = pop32 ();
push_type (t2);
push_type (t1);
push_type (t4);
}
push_type (t3);
push_type (t2);
push_type (t1);
}
}
break;
case op_swap:
{
type t1 = pop32 ();
type t2 = pop32 ();
push_type (t1);
push_type (t2);
}
break;
case op_iadd:
case op_isub:
case op_imul:
case op_idiv:
case op_irem:
case op_ishl:
case op_ishr:
case op_iushr:
case op_iand:
case op_ior:
case op_ixor:
pop_type (int_type);
push_type (pop_type (int_type));
break;
case op_ladd:
case op_lsub:
case op_lmul:
case op_ldiv:
case op_lrem:
case op_land:
case op_lor:
case op_lxor:
pop_type (long_type);
push_type (pop_type (long_type));
break;
case op_lshl:
case op_lshr:
case op_lushr:
pop_type (int_type);
push_type (pop_type (long_type));
break;
case op_fadd:
case op_fsub:
case op_fmul:
case op_fdiv:
case op_frem:
pop_type (float_type);
push_type (pop_type (float_type));
break;
case op_dadd:
case op_dsub:
case op_dmul:
case op_ddiv:
case op_drem:
pop_type (double_type);
push_type (pop_type (double_type));
break;
case op_ineg:
case op_i2b:
case op_i2c:
case op_i2s:
push_type (pop_type (int_type));
break;
case op_lneg:
push_type (pop_type (long_type));
break;
case op_fneg:
push_type (pop_type (float_type));
break;
case op_dneg:
push_type (pop_type (double_type));
break;
case op_iinc:
get_variable (get_byte (), int_type);
get_byte ();
break;
case op_i2l:
pop_type (int_type);
push_type (long_type);
break;
case op_i2f:
pop_type (int_type);
push_type (float_type);
break;
case op_i2d:
pop_type (int_type);
push_type (double_type);
break;
case op_l2i:
pop_type (long_type);
push_type (int_type);
break;
case op_l2f:
pop_type (long_type);
push_type (float_type);
break;
case op_l2d:
pop_type (long_type);
push_type (double_type);
break;
case op_f2i:
pop_type (float_type);
push_type (int_type);
break;
case op_f2l:
pop_type (float_type);
push_type (long_type);
break;
case op_f2d:
pop_type (float_type);
push_type (double_type);
break;
case op_d2i:
pop_type (double_type);
push_type (int_type);
break;
case op_d2l:
pop_type (double_type);
push_type (long_type);
break;
case op_d2f:
pop_type (double_type);
push_type (float_type);
break;
case op_lcmp:
pop_type (long_type);
pop_type (long_type);
push_type (int_type);
break;
case op_fcmpl:
case op_fcmpg:
pop_type (float_type);
pop_type (float_type);
push_type (int_type);
break;
case op_dcmpl:
case op_dcmpg:
pop_type (double_type);
pop_type (double_type);
push_type (int_type);
break;
case op_ifeq:
case op_ifne:
case op_iflt:
case op_ifge:
case op_ifgt:
case op_ifle:
pop_type (int_type);
push_jump (get_short ());
break;
case op_if_icmpeq:
case op_if_icmpne:
case op_if_icmplt:
case op_if_icmpge:
case op_if_icmpgt:
case op_if_icmple:
pop_type (int_type);
pop_type (int_type);
push_jump (get_short ());
break;
case op_if_acmpeq:
case op_if_acmpne:
pop_type (reference_type);
pop_type (reference_type);
push_jump (get_short ());
break;
case op_goto:
push_jump (get_short ());
invalidate_pc ();
break;
case op_jsr:
handle_jsr_insn (get_short ());
break;
case op_ret:
handle_ret_insn (get_byte ());
break;
case op_tableswitch:
{
pop_type (int_type);
skip_padding ();
push_jump (get_int ());
jint low = get_int ();
jint high = get_int ();
// Already checked LOW -vs- HIGH.
for (int i = low; i <= high; ++i)
push_jump (get_int ());
invalidate_pc ();
}
break;
case op_lookupswitch:
{
pop_type (int_type);
skip_padding ();
push_jump (get_int ());
jint npairs = get_int ();
// Already checked NPAIRS >= 0.
jint lastkey = 0;
for (int i = 0; i < npairs; ++i)
{
jint key = get_int ();
if (i > 0 && key <= lastkey)
verify_fail ("lookupswitch pairs unsorted", start_PC);
lastkey = key;
push_jump (get_int ());
}
invalidate_pc ();
}
break;
case op_ireturn:
check_return_type (pop_type (int_type));
invalidate_pc ();
break;
case op_lreturn:
check_return_type (pop_type (long_type));
invalidate_pc ();
break;
case op_freturn:
check_return_type (pop_type (float_type));
invalidate_pc ();
break;
case op_dreturn:
check_return_type (pop_type (double_type));
invalidate_pc ();
break;
case op_areturn:
check_return_type (pop_init_ref (reference_type));
invalidate_pc ();
break;
case op_return:
// We only need to check this when the return type is
// void, because all instance initializers return void.
if (this_is_init)
current_state->check_this_initialized (this);
check_return_type (void_type);
invalidate_pc ();
break;
case op_getstatic:
push_type (check_field_constant (get_ushort ()));
break;
case op_putstatic:
pop_type (check_field_constant (get_ushort ()));
break;
case op_getfield:
{
type klass;
type field = check_field_constant (get_ushort (), &klass);
pop_type (klass);
push_type (field);
}
break;
case op_putfield:
{
type klass;
type field = check_field_constant (get_ushort (), &klass);
pop_type (field);
// We have an obscure special case here: we can use
// `putfield' on a field declared in this class, even if
// `this' has not yet been initialized.
if (! current_state->this_type.isinitialized ()
&& current_state->this_type.pc == type::SELF)
klass.set_uninitialized (type::SELF, this);
pop_type (klass);
}
break;
case op_invokevirtual:
case op_invokespecial:
case op_invokestatic:
case op_invokeinterface:
{
_Jv_Utf8Const *method_name, *method_signature;
type class_type
= check_method_constant (get_ushort (),
opcode == op_invokeinterface,
&method_name,
&method_signature);
// NARGS is only used when we're processing
// invokeinterface. It is simplest for us to compute it
// here and then verify it later.
int nargs = 0;
if (opcode == op_invokeinterface)
{
nargs = get_byte ();
if (get_byte () != 0)
verify_fail ("invokeinterface dummy byte is wrong");
}
bool is_init = false;
if (_Jv_equalUtf8Consts (method_name, gcj::init_name))
{
is_init = true;
if (opcode != op_invokespecial)
verify_fail ("can't invoke <init>");
}
else if (method_name->data[0] == '<')
verify_fail ("can't invoke method starting with `<'");
// Pop arguments and check types.
int arg_count = _Jv_count_arguments (method_signature);
type arg_types[arg_count];
compute_argument_types (method_signature, arg_types);
for (int i = arg_count - 1; i >= 0; --i)
{
// This is only used for verifying the byte for
// invokeinterface.
nargs -= arg_types[i].depth ();
pop_init_ref (arg_types[i]);
}
if (opcode == op_invokeinterface
&& nargs != 1)
verify_fail ("wrong argument count for invokeinterface");
if (opcode != op_invokestatic)
{
type t = class_type;
if (is_init)
{
// In this case the PC doesn't matter.
t.set_uninitialized (type::UNINIT, this);
// FIXME: check to make sure that the <init>
// call is to the right class.
// It must either be super or an exact class
// match.
}
type raw = pop_raw ();
if (! t.compatible (raw, this))
verify_fail ("incompatible type on stack");
if (is_init)
current_state->set_initialized (raw.get_pc (),
current_method->max_locals);
}
type rt = compute_return_type (method_signature);
if (! rt.isvoid ())
push_type (rt);
}
break;
case op_new:
{
type t = check_class_constant (get_ushort ());
if (t.isarray () || t.isinterface (this) || t.isabstract (this))
verify_fail ("type is array, interface, or abstract");
t.set_uninitialized (start_PC, this);
push_type (t);
}
break;
case op_newarray:
{
int atype = get_byte ();
// We intentionally have chosen constants to make this
// valid.
if (atype < boolean_type || atype > long_type)
verify_fail ("type not primitive", start_PC);
pop_type (int_type);
type t (construct_primitive_array_type (type_val (atype)), this);
push_type (t);
}
break;
case op_anewarray:
pop_type (int_type);
push_type (check_class_constant (get_ushort ()).to_array (this));
break;
case op_arraylength:
{
type t = pop_init_ref (reference_type);
if (! t.isarray () && ! t.isnull ())
verify_fail ("array type expected");
push_type (int_type);
}
break;
case op_athrow:
pop_type (type (&java::lang::Throwable::class$, this));
invalidate_pc ();
break;
case op_checkcast:
pop_init_ref (reference_type);
push_type (check_class_constant (get_ushort ()));
break;
case op_instanceof:
pop_init_ref (reference_type);
check_class_constant (get_ushort ());
push_type (int_type);
break;
case op_monitorenter:
pop_init_ref (reference_type);
break;
case op_monitorexit:
pop_init_ref (reference_type);
break;
case op_wide:
{
switch (get_byte ())
{
case op_iload:
push_type (get_variable (get_ushort (), int_type));
break;
case op_lload:
push_type (get_variable (get_ushort (), long_type));
break;
case op_fload:
push_type (get_variable (get_ushort (), float_type));
break;
case op_dload:
push_type (get_variable (get_ushort (), double_type));
break;
case op_aload:
push_type (get_variable (get_ushort (), reference_type));
break;
case op_istore:
set_variable (get_ushort (), pop_type (int_type));
break;
case op_lstore:
set_variable (get_ushort (), pop_type (long_type));
break;
case op_fstore:
set_variable (get_ushort (), pop_type (float_type));
break;
case op_dstore:
set_variable (get_ushort (), pop_type (double_type));
break;
case op_astore:
set_variable (get_ushort (), pop_init_ref (reference_type));
break;
case op_ret:
handle_ret_insn (get_short ());
break;
case op_iinc:
get_variable (get_ushort (), int_type);
get_short ();
break;
default:
verify_fail ("unrecognized wide instruction", start_PC);
}
}
break;
case op_multianewarray:
{
type atype = check_class_constant (get_ushort ());
int dim = get_byte ();
if (dim < 1)
verify_fail ("too few dimensions to multianewarray", start_PC);
atype.verify_dimensions (dim, this);
for (int i = 0; i < dim; ++i)
pop_type (int_type);
push_type (atype);
}
break;
case op_ifnull:
case op_ifnonnull:
pop_type (reference_type);
push_jump (get_short ());
break;
case op_goto_w:
push_jump (get_int ());
invalidate_pc ();
break;
case op_jsr_w:
handle_jsr_insn (get_int ());
break;
// These are unused here, but we call them out explicitly
// so that -Wswitch-enum doesn't complain.
case op_putfield_1:
case op_putfield_2:
case op_putfield_4:
case op_putfield_8:
case op_putfield_a:
case op_putstatic_1:
case op_putstatic_2:
case op_putstatic_4:
case op_putstatic_8:
case op_putstatic_a:
case op_getfield_1:
case op_getfield_2s:
case op_getfield_2u:
case op_getfield_4:
case op_getfield_8:
case op_getfield_a:
case op_getstatic_1:
case op_getstatic_2s:
case op_getstatic_2u:
case op_getstatic_4:
case op_getstatic_8:
case op_getstatic_a:
default:
// Unrecognized opcode.
verify_fail ("unrecognized instruction in verify_instructions_0",
start_PC);
}
}
}
public:
void verify_instructions ()
{
branch_prepass ();
verify_instructions_0 ();
}
_Jv_BytecodeVerifier (_Jv_InterpMethod *m)
{
// We just print the text as utf-8. This is just for debugging
// anyway.
debug_print ("--------------------------------\n");
debug_print ("-- Verifying method `%s'\n", m->self->name->data);
current_method = m;
bytecode = m->bytecode ();
exception = m->exceptions ();
current_class = m->defining_class;
states = NULL;
flags = NULL;
jsr_ptrs = NULL;
utf8_list = NULL;
isect_list = NULL;
entry_points = NULL;
}
~_Jv_BytecodeVerifier ()
{
if (states)
_Jv_Free (states);
if (flags)
_Jv_Free (flags);
if (jsr_ptrs)
{
for (int i = 0; i < current_method->code_length; ++i)
{
if (jsr_ptrs[i] != NULL)
{
subr_info *info = jsr_ptrs[i];
while (info != NULL)
{
subr_info *next = info->next;
_Jv_Free (info);
info = next;
}
}
}
_Jv_Free (jsr_ptrs);
}
while (utf8_list != NULL)
{
linked_utf8 *n = utf8_list->next;
_Jv_Free (utf8_list->val);
_Jv_Free (utf8_list);
utf8_list = n;
}
while (entry_points != NULL)
{
subr_entry_info *next = entry_points->next;
_Jv_Free (entry_points);
entry_points = next;
}
while (isect_list != NULL)
{
ref_intersection *next = isect_list->alloc_next;
delete isect_list;
isect_list = next;
}
}
};
void
_Jv_VerifyMethod (_Jv_InterpMethod *meth)
{
_Jv_BytecodeVerifier v (meth);
v.verify_instructions ();
}
#endif /* INTERPRETER */