blob: 945a5354124e012ad73337d8e66a04abfae158b0 [file] [log] [blame]
/*
* Copyright (c) 2004, 2015, Oracle and/or its affiliates. All rights reserved.
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
*
* This code is free software; you can redistribute it and/or modify it
* under the terms of the GNU General Public License version 2 only, as
* published by the Free Software Foundation.
*
* This code is distributed in the hope that it will be useful, but WITHOUT
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
* version 2 for more details (a copy is included in the LICENSE file that
* accompanied this code).
*
* You should have received a copy of the GNU General Public License version
* 2 along with this work; if not, write to the Free Software Foundation,
* Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
*
* Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
* or visit www.oracle.com if you need additional information or have any
* questions.
*/
/**
* @test
* @bug 5033550
* @summary JDWP back end uses modified UTF-8
* @author jjh
*
* @run build TestScaffold VMConnection TargetListener TargetAdapter
* @run compile -g UTF8Test.java
* @run driver UTF8Test
*/
/*
There is UTF-8 and there is modified UTF-8, which I will call M-UTF-8.
The two differ in the representation of binary 0, and
in some other more esoteric representations.
See
http://java.sun.com/developer/technicalArticles/Intl/Supplementary/#Modified_UTF-8
http://java.sun.com/javase/6/docs/technotes/guides/jni/spec/types.html#wp16542
All the following are observations of the treatment
of binary 0. In UTF-8, this represented as one byte:
0x00
while in modified UTF-8, it is represented as two bytes
0xc0 0x80
** I haven't investigated if the other differences between UTF-8 and
M-UTF-8 are handled in the same way.
Here is how these our handled in our BE, JDWP, and FE:
- Strings in .class files are M-UTF-8.
- To get the value of a string object from the VM, our BE calls
char * utf = JNI_FUNC_PTR(env,GetStringUTFChars)(env, string, NULL);
which returns M-UTF-8.
- To create a string object in the VM, our BE VirtualMachine.createString() calls
string = JNI_FUNC_PTR(env,NewStringUTF)(env, cstring);
This function expects the string to be M-UTF-8
BUG: If the string came from JDWP, then it is actually UTF-8
- I haven't investigated strings in JVMTI.
- The JDWP spec says that strings are UTF-8. The intro
says this for all strings, and the createString command and
the StringRefernce.value command say it explicitly.
- Our FE java writes strings to JDWP as UTF-8.
- BE function outStream_writeString uses strlen meaning
it expects no 0 bytes, meaning that it expects M-UTF-8
This function writes the byte length and then calls
outStream.c::writeBytes which just writes the bytes to JDWP as is.
BUG: If such a string came from the VM via JNI, it is actually
M-UTF-8
FIX: - scan string to see if contains an M-UTF-8 char.
if yes,
- call String(bytes, 0, len, "UTF8")
to get a java string. Will this work -ie, the
input is M-UTF-8 instead of real UTF-8
- call some java method (NOT JNI which
would just come back with M-UTF-8)
on the String to get real UTF-8
- The JDWP StringReference.value command does reads a string
from the BE out of the JDWP stream and does this to
createe a Java String for it (see PacketStream.readString):
String readString() {
String ret;
int len = readInt();
try {
ret = new String(pkt.data, inCursor, len, "UTF8");
} catch(java.io.UnsupportedEncodingException e) {
This String ctor converts _both- the M-UTF-8 0xc0 0x80
and UTF-8 0x00 into a Java char containing 0x0000
Does it do this for the other differences too?
Summary:
1. JDWP says strings are UTF-8.
We interpret this to mean standard UTF-8.
2. JVMTI will be changed to match JNI saying that strings
are M-UTF-8.
3. The BE gets UTF-8 strings off JDWP and must convert them to
M-UTF-8 before giving it to JVMTI or JNI.
4. The BE gets M-UTF-8 strings from JNI and JVMTI and
must convert them to UTF-8 when writing to JDWP.
Here is how the supplementals are represented in java Strings.
This from java.lang.Character doc:
The Java 2 platform uses the UTF-16 representation in char arrays and
in the String and StringBuffer classes. In this representation,
supplementary characters are represented as a pair of char values,
the first from the high-surrogates range, (\uD800-\uDBFF), the second
from the low-surrogates range (\uDC00-\uDFFF).
See utf8.txt
----
NSK Packet.java in the nsk/share/jdwp framework does this to write
a string to JDWP:
public void addString(String value) {
final int count = JDWP.TypeSize.INT + value.length();
addInt(value.length());
try {
addBytes(value.getBytes("UTF-8"), 0, value.length());
} catch (UnsupportedEncodingException e) {
throw new Failure("Unsupported UTF-8 ecnoding while adding string value to JDWP packet:\n\t"
+ e);
}
}
?? Does this get the standard UTF-8? I would expect so.
and the readString method does this:
for (int i = 0; i < len; i++)
s[i] = getByte();
try {
return new String(s, "UTF-8");
} catch (UnsupportedEncodingException e) {
throw new Failure("Unsupported UTF-8 ecnoding while extracting string value from JDWP packet:\n\t"
+ e);
}
Thus, this won't notice the modified UTF-8 coming in from JDWP .
*/
import com.sun.jdi.*;
import com.sun.jdi.event.*;
import com.sun.jdi.request.*;
import java.io.UnsupportedEncodingException;
import java.util.*;
/********** target program **********/
/*
* The debuggee has a few Strings the debugger reads via JDI
*/
class UTF8Targ {
static String[] vals = new String[] {"xx\u0000yy", // standard UTF-8 0
"xx\ud800\udc00yy", // first supplementary
"xx\udbff\udfffyy" // last supplementary
// d800 = 1101 1000 0000 0000 dc00 = 1101 1100 0000 0000
// dbff = 1101 1011 1111 1111 dfff = 1101 1111 1111 1111
};
static String aField;
public static void main(String[] args){
System.out.println("Howdy!");
gus();
System.out.println("Goodbye from UTF8Targ!");
}
static void gus() {
}
}
/********** test program **********/
public class UTF8Test extends TestScaffold {
ClassType targetClass;
ThreadReference mainThread;
Field targetField;
UTF8Test (String args[]) {
super(args);
}
public static void main(String[] args) throws Exception {
new UTF8Test(args).startTests();
}
/********** test core **********/
protected void runTests() throws Exception {
/*
* Get to the top of main()
* to determine targetClass and mainThread
*/
BreakpointEvent bpe = startToMain("UTF8Targ");
targetClass = (ClassType)bpe.location().declaringType();
targetField = targetClass.fieldByName("aField");
ArrayReference targetVals = (ArrayReference)targetClass.getValue(targetClass.fieldByName("vals"));
/* For each string in the debuggee's 'val' array, verify that we can
* read that value via JDI.
*/
for (int ii = 0; ii < UTF8Targ.vals.length; ii++) {
StringReference val = (StringReference)targetVals.getValue(ii);
String valStr = val.value();
/*
* Verify that we can read a value correctly.
* We read it via JDI, and access it directly from the static
* var in the debuggee class.
*/
if (!valStr.equals(UTF8Targ.vals[ii]) ||
valStr.length() != UTF8Targ.vals[ii].length()) {
failure(" FAILED: Expected /" + printIt(UTF8Targ.vals[ii]) +
"/, but got /" + printIt(valStr) + "/, length = " + valStr.length());
}
}
/* Test 'all' unicode chars - send them to the debuggee via JDI
* and then read them back.
*/
doFancyVersion();
resumeTo("UTF8Targ", "gus", "()V");
try {
Thread.sleep(1000);
} catch (InterruptedException ee) {
}
/*
* resume the target listening for events
*/
listenUntilVMDisconnect();
/*
* deal with results of test
* if anything has called failure("foo") testFailed will be true
*/
if (!testFailed) {
println("UTF8Test: passed");
} else {
throw new Exception("UTF8Test: failed");
}
}
/**
* For each unicode value, send a string containing
* it to the debuggee via JDI, read it back via JDI, and see if
* we get the same value.
*/
void doFancyVersion() throws Exception {
// This does 4 chars at a time just to save time.
for (int ii = Character.MIN_CODE_POINT;
ii < Character.MIN_SUPPLEMENTARY_CODE_POINT;
ii += 4) {
// Skip the surrogates
if (ii == Character.MIN_SURROGATE) {
ii = Character.MAX_SURROGATE - 3;
break;
}
doFancyTest(ii, ii + 1, ii + 2, ii + 3);
}
// Do the supplemental chars.
for (int ii = Character.MIN_SUPPLEMENTARY_CODE_POINT;
ii <= Character.MAX_CODE_POINT;
ii += 2000) {
// Too many of these so just do a few
doFancyTest(ii, ii + 1, ii + 2, ii + 3);
}
}
void doFancyTest(int ... args) throws Exception {
String ss = new String(args, 0, 4);
targetClass.setValue(targetField, vm().mirrorOf(ss));
StringReference returnedVal = (StringReference)targetClass.getValue(targetField);
String returnedStr = returnedVal.value();
if (!ss.equals(returnedStr)) {
failure("Set: FAILED: Expected /" + printIt(ss) +
"/, but got /" + printIt(returnedStr) + "/, length = " + returnedStr.length());
}
}
/**
* Return a String containing binary representations of
* the chars in a String.
*/
String printIt(String arg) {
char[] carray = arg.toCharArray();
StringBuffer bb = new StringBuffer(arg.length() * 5);
for (int ii = 0; ii < arg.length(); ii++) {
int ccc = arg.charAt(ii);
bb.append(String.format("%1$04x ", ccc));
}
return bb.toString();
}
String printIt1(String arg) {
byte[] barray = null;
try {
barray = arg.getBytes("UTF-8");
} catch (UnsupportedEncodingException ee) {
}
StringBuffer bb = new StringBuffer(barray.length * 3);
for (int ii = 0; ii < barray.length; ii++) {
bb.append(String.format("%1$02x ", barray[ii]));
}
return bb.toString();
}
}