/*
 * Copyright (c) 2010, 2017 Oracle and/or its affiliates. All rights reserved.
 * Copyright 2004 The Apache Software Foundation
 *
 * Licensed under the Apache License, Version 2.0 (the "License");
 * you may not use this file except in compliance with the License.
 * You may obtain a copy of the License at
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */

package org.glassfish.grizzly.http.util;

import java.io.CharConversionException;
import java.io.IOException;
import java.nio.charset.Charset;
import java.util.ArrayList;
import java.util.Collections;
import java.util.LinkedHashMap;
import java.util.Set;
import java.util.logging.Level;
import java.util.logging.Logger;
import org.glassfish.grizzly.Buffer;

import org.glassfish.grizzly.Grizzly;

import org.glassfish.grizzly.localization.LogMessages;

Author:Costin Manolache
/** * @author Costin Manolache */
public final class Parameters {
Default Logger.
/** * Default Logger. */
private final static Logger LOGGER = Grizzly.logger(Parameters.class); private final LinkedHashMap<String, ArrayList<String>> paramHashValues = new LinkedHashMap<String, ArrayList<String>>(); private boolean didQueryParameters = false; private boolean didMerge = false; MimeHeaders headers; DataChunk queryDC; final DataChunk decodedQuery = DataChunk.newInstance(); public static final int INITIAL_SIZE = 4; // Garbage-less parameter merging. // In a sub-request with parameters, the new parameters // will be stored in child. When a getParameter happens, // the 2 are merged together. The child will be altered // to contain the merged values - the parent is always the // original request. private Parameters child = null; private Parameters parent = null; private Parameters currentChild = null; Charset encoding = null; Charset queryStringEncoding = null; private int limit = -1; private int parameterCount = 0; public void setQuery(final DataChunk queryBC) { this.queryDC = queryBC; } public void setHeaders(final MimeHeaders headers) { this.headers = headers; } public void setLimit(int limit) { this.limit = limit; } public void setEncoding(final Charset encoding) { this.encoding = encoding; if (LOGGER.isLoggable(Level.FINEST)) { LOGGER.log(Level.FINEST, "Set encoding to {0}", encoding); } } public Charset getEncoding() { return encoding; } public void setQueryStringEncoding(final Charset queryStringEncoding) { this.queryStringEncoding = queryStringEncoding; if (LOGGER.isLoggable(Level.FINEST)) { LOGGER.log(Level.FINEST, "Set query string encoding to {0}", queryStringEncoding); } } public Charset getQueryStringEncoding() { return queryStringEncoding; } public void recycle() { paramHashValues.clear(); didQueryParameters = false; currentChild = null; didMerge = false; encoding = null; queryStringEncoding = null; parameterCount = 0; decodedQuery.recycle(); } // -------------------- Sub-request support -------------------- public Parameters getCurrentSet() { if (currentChild == null) { return this; } return currentChild; }
Create ( or reuse ) a child that will be used during a sub-request. All future changes ( setting query string, adding parameters ) will affect the child ( the parent request is never changed ). Both setters and getters will return the data from the deepest child, merged with data from parents.
/** * Create ( or reuse ) a child that will be used during a sub-request. All future changes ( setting query string, * adding parameters ) will affect the child ( the parent request is never changed ). Both setters and getters will * return the data from the deepest child, merged with data from parents. */
public void push() { // We maintain a linked list, that will grow to the size of the // longest include chain. // The list has 2 points of interest: // - request.parameters() is the original request and head, // - request.parameters().currentChild() is the current set. // The ->child and parent<- links are preserved ( currentChild is not // the last in the list ) // create a new element in the linked list // note that we reuse the child, if any - pop will not // set child to null ! if (currentChild == null) { currentChild = new Parameters(); // currentChild.setURLDecoder(urlDec); currentChild.parent = this; return; } if (currentChild.child == null) { currentChild.child = new Parameters(); // currentChild.setURLDecoder(urlDec); currentChild.child.parent = currentChild; } // it is not null if this object already had a child // i.e. a deeper include() ( we keep it ) // the head will be the new element. currentChild = currentChild.child; currentChild.setEncoding(encoding); }
Discard the last child. This happens when we return from a sub-request and the parameters are locally modified.
/** * Discard the last child. This happens when we return from a sub-request and the parameters are locally modified. */
public void pop() { if (currentChild == null) { throw new RuntimeException("Attempt to pop without a push"); } currentChild.recycle(); currentChild = currentChild.parent; // don't remove the top. } // -------------------- Data access -------------------- // Access to the current name/values, no side effect ( processing ). // You must explicitely call handleQueryParameters and the post methods. // This is the original data representation ( hash of String->String[]) public void addParameterValues(String key, String[] newValues) { if (key == null) { return; } ArrayList<String> values; if (paramHashValues.containsKey(key)) { values = paramHashValues.get(key); } else { values = new ArrayList<String>(1); paramHashValues.put(key, values); } values.ensureCapacity(values.size() + newValues.length); Collections.addAll(values, newValues); } public String[] getParameterValues(String name) { handleQueryParameters(); final ArrayList<String> values; // sub-request if (currentChild != null) { currentChild.merge(); values = currentChild.paramHashValues.get(name); } else { // no "facade" values = paramHashValues.get(name); } return ((values != null) ? values.toArray(new String[values.size()]) : null); } public Set<String> getParameterNames() { handleQueryParameters(); // Slow - the original code if (currentChild != null) { currentChild.merge(); /* START PWC 6057385 return currentChild.paramHashStringArray.keys(); */ // START PWC 6057385 currentChild.paramHashValues.keySet(); // END PWC 6057385 } // merge in child /* START PWC 6057385 return paramHashStringArray.keys(); */ // START PWC 6057385 return paramHashValues.keySet(); // END PWC 6057385 }
Combine the parameters from parent with our local ones
/** * Combine the parameters from parent with our local ones */
private void merge() { // recursive if (LOGGER.isLoggable(Level.FINEST)) { LOGGER.log(Level.FINEST, "Before merging {0} {1} {2}", new Object[]{this, parent, didMerge}); LOGGER.log(Level.FINEST, paramsAsString()); } // Local parameters first - they take precedence as in spec. handleQueryParameters(); // we already merged with the parent if (didMerge) { return; } // we are the top level if (parent == null) { return; } // Add the parent props to the child ( lower precedence ) parent.merge(); /* START PWC 6057385 Hashtable parentProps=parent.paramHashStringArray; */ // START PWC 6057385 LinkedHashMap<String, ArrayList<String>> parentProps = parent.paramHashValues; // END PWC 6057385 merge2(paramHashValues, parentProps); didMerge = true; if (LOGGER.isLoggable(Level.FINEST)) { LOGGER.log(Level.FINEST, "After {0}", paramsAsString()); } } // Shortcut. public String getParameter(final String name) { ArrayList<String> values = paramHashValues.get(name); if (values != null) { if (values.isEmpty()) { return ""; } return values.get(0); } else { return null; } } // -------------------- Processing --------------------
Process the query string into parameters
/** * Process the query string into parameters */
public void handleQueryParameters() { if (didQueryParameters) { return; } didQueryParameters = true; if (queryDC == null || queryDC.isNull()) { return; } if (LOGGER.isLoggable(Level.FINEST)) { LOGGER.log(Level.FINEST, "Decoding query {0} {1}", new Object[]{queryDC, queryStringEncoding}); } decodedQuery.duplicate(queryDC); processParameters(decodedQuery, queryStringEncoding); } // --------------------
Combine 2 hashtables into a new one. ( two will be added to one ). Used to combine child parameters ( RequestDispatcher's query ) with parent parameters ( original query or parent dispatcher )
/** * Combine 2 hashtables into a new one. ( two will be added to one ). Used to combine child parameters ( * RequestDispatcher's query ) with parent parameters ( original query or parent dispatcher ) */
/* START PWC 6057385 private static void merge2(Hashtable one, Hashtable two ) { Enumeration e = two.keys(); while (e.hasMoreElements()) { String name = (String) e.nextElement(); */ // START PWC 6057385 private static void merge2(LinkedHashMap<String, ArrayList<String>> one, LinkedHashMap<String, ArrayList<String>> two) { for (String name : two.keySet()) { // END PWC 6057385 ArrayList<String> oneValue = one.get(name); ArrayList<String> twoValue = two.get(name); ArrayList<String> combinedValue; if (twoValue != null) { if (oneValue == null) { combinedValue = new ArrayList<String>(twoValue); } else { combinedValue = new ArrayList<String>(oneValue.size() + twoValue.size()); combinedValue.addAll(oneValue); combinedValue.addAll(twoValue); } one.put(name, combinedValue); } } } public void addParameter(String key, String value) throws IllegalStateException { if (key == null) { return; } parameterCount++; if (limit > -1 && parameterCount > limit) { // Processing this parameter will push us over the limit. ISE is // what Request.parseParts() uses for requests that are too big throw new IllegalStateException(//sm.getString( //"parameters.maxCountFail", Integer.valueOf(limit))); ); } ArrayList<String> values = paramHashValues.get(key); if (values == null) { values = new ArrayList<String>(1); paramHashValues.put(key, values); } values.add(value); } // public void setURLDecoder(UDecoder u) { // urlDec = u; // } // -------------------- Parameter parsing -------------------- // This code is not used right now - it's the optimized version // of the above. // we are called from a single thread - we can do it the hard way // if needed final BufferChunk tmpName = new BufferChunk(); final BufferChunk tmpValue = new BufferChunk(); private final BufferChunk origName = new BufferChunk(); private final BufferChunk origValue = new BufferChunk(); final CharChunk tmpNameC = new CharChunk(1024); final CharChunk tmpValueC = new CharChunk(1024); public static final String DEFAULT_ENCODING = Constants.DEFAULT_HTTP_CHARACTER_ENCODING; public static final Charset DEFAULT_CHARSET = Constants.DEFAULT_HTTP_CHARSET; public void processParameters(final Buffer buffer, final int start, final int len) { processParameters(buffer, start, len, encoding); } public void processParameters(final Buffer buffer, final int start, final int len, final Charset enc) { if (LOGGER.isLoggable(Level.FINEST)) { LOGGER.log(Level.FINEST, "Process parameters. Buffer: {0} start={1} len={2} content={3}", new Object[]{ buffer, start, len, buffer.toStringContent(enc, start, start + len) }); } int decodeFailCount = 0; int end = start + len; int pos = start; while (pos < end) { if (limit > -1 && parameterCount >= limit) { LOGGER.warning(LogMessages.WARNING_GRIZZLY_HTTP_SEVERE_GRIZZLY_HTTP_PARAMETERS_MAX_COUNT_FAIL(limit)); break; } int nameStart = pos; int nameEnd = -1; int valueStart = -1; int valueEnd = -1; boolean parsingName = true; boolean decodeName = false; boolean decodeValue = false; boolean parameterComplete = false; do { switch (buffer.get(pos)) { case '=': if (parsingName) { // Name finished. Value starts from next character nameEnd = pos; parsingName = false; valueStart = ++pos; } else { // Equals character in value pos++; } break; case '&': if (parsingName) { // Name finished. No value. nameEnd = pos; } else { // Value finished valueEnd = pos; } parameterComplete = true; pos++; break; case '+': case '%': // Decoding required if (parsingName) { decodeName = true; } else { decodeValue = true; } pos++; break; default: pos++; break; } } while (!parameterComplete && pos < end); if (pos == end) { if (nameEnd == -1) { nameEnd = pos; } else if (valueStart > -1 && valueEnd == -1) { valueEnd = pos; } } if (LOGGER.isLoggable(Level.FINEST) && valueStart == -1) { LOGGER.log(Level.FINEST, LogMessages.FINE_GRIZZLY_HTTP_PARAMETERS_NOEQUAL( nameStart, nameEnd, buffer.toStringContent(DEFAULT_CHARSET, nameStart, nameEnd))); } if (nameEnd <= nameStart) { if (LOGGER.isLoggable(Level.INFO)) { String extract; if (valueEnd < nameStart) { LOGGER.info(LogMessages.INFO_GRIZZLY_HTTP_PARAMETERS_INVALID_CHUNK( nameStart, nameEnd, null)); } } continue; // invalid chunk - it's better to ignore } tmpName.setBufferChunk(buffer, nameStart, nameEnd); tmpValue.setBufferChunk(buffer, valueStart, valueEnd); // Take copies as if anything goes wrong originals will be // corrupted. This means original values can be logged. // For performance - only done for debug if (LOGGER.isLoggable(Level.FINEST)) { origName.setBufferChunk(buffer, nameStart, nameEnd); origValue.setBufferChunk(buffer, valueStart, valueEnd); } try { String name; String value; if (decodeName) { name = urlDecode(tmpName, enc); } else { name = tmpName.toString(enc); } if (valueStart != -1) { if (decodeValue) { value = urlDecode(tmpValue, enc); } else { value = tmpValue.toString(enc); } } else { value = ""; } addParameter(name, value); } catch (Exception e) { decodeFailCount++; if (LOGGER.isLoggable(Level.FINEST)) { LOGGER.log(Level.FINEST, LogMessages.FINE_GRIZZLY_HTTP_PARAMETERS_DECODE_FAIL_DEBUG( origName.toString(), origValue.toString())); } else if (LOGGER.isLoggable(Level.INFO) && decodeFailCount == 1) { final String name = ((tmpName.getLength() > 0) ? tmpName.toString() : "unavailable"); final String value = ((tmpValue.getLength() > 0) ? tmpValue.toString() : "unavailable"); LOGGER.log(Level.INFO, LogMessages.INFO_GRIZZLY_HTTP_PARAMETERS_DECODE_FAIL_INFO( e.getMessage(), name, value)); LOGGER.log(Level.FINE, "Decoding stacktrace.", e); } } finally { tmpName.recycle(); tmpValue.recycle(); } } if (!LOGGER.isLoggable(Level.FINEST) && decodeFailCount > 1) { LOGGER.info(LogMessages.INFO_GRIZZLY_HTTP_PARAMETERS_MULTIPLE_DECODING_FAIL(decodeFailCount)); } } private String urlDecode(final BufferChunk bc, final Charset enc) throws IOException { // if (urlDec == null) { // urlDec = new UDecoder(); // } URLDecoder.decode(bc, true); String result; if (enc != null) { if (bc.getStart() == -1 && bc.getEnd() == -1) { return ""; } result = bc.toString(enc); } else { final CharChunk cc = tmpNameC; final int length = bc.getLength(); cc.allocate(length, -1); // Default encoding: fast conversion final Buffer bbuf = bc.getBuffer(); final char[] cbuf = cc.getBuffer(); final int start = bc.getStart(); for (int i = 0; i < length; i++) { cbuf[i] = (char) (bbuf.get(i + start) & 0xff); } cc.setChars(cbuf, 0, length); result = cc.toString(); cc.recycle(); } return result; } public void processParameters(char chars[], int start, int len) { int end = start + len; int pos = start; int decodeFailCount = 0; if (LOGGER.isLoggable(Level.FINEST)) { LOGGER.log(Level.FINEST, "Process parameters. chars: {0} start={1} len={2} content={3}", new Object[]{ chars, start, len, new String(chars, start, len) }); } do { if (limit > -1 && parameterCount >= limit) { LOGGER.warning( LogMessages.WARNING_GRIZZLY_HTTP_SEVERE_GRIZZLY_HTTP_PARAMETERS_MAX_COUNT_FAIL( limit)); break; } boolean noEq = false; int nameStart = pos; int valStart = -1; int valEnd = -1; int nameEnd = CharChunk.indexOf(chars, nameStart, end, '='); int nameEnd2 = CharChunk.indexOf(chars, nameStart, end, '&'); if ((nameEnd2 != -1) && (nameEnd == -1 || nameEnd > nameEnd2)) { nameEnd = nameEnd2; noEq = true; valStart = nameEnd; valEnd = nameEnd; if (LOGGER.isLoggable(Level.FINEST)) { LOGGER.log(Level.FINEST, "no equal {0} {1} {2}", new Object[]{ nameStart, nameEnd, new String(chars, nameStart, nameEnd - nameStart) }); } } if (nameEnd == -1) { nameEnd = end; } if (!noEq) { valStart = (nameEnd < end) ? nameEnd + 1 : end; valEnd = CharChunk.indexOf(chars, valStart, end, '&'); if (valEnd == -1) { valEnd = (valStart < end) ? end : valStart; } } pos = valEnd + 1; if (nameEnd <= nameStart) { continue; // invalid chunk - no name, it's better to ignore // XXX log it ? } try { tmpNameC.append(chars, nameStart, nameEnd - nameStart); tmpValueC.append(chars, valStart, valEnd - valStart); if (LOGGER.isLoggable(Level.FINEST)) { LOGGER.log(Level.FINEST, "{0}= {1}", new Object[]{tmpNameC, tmpValueC}); } URLDecoder.decode(tmpNameC, tmpNameC, true, queryStringEncoding.name()); URLDecoder.decode(tmpValueC, tmpValueC, true, queryStringEncoding.name()); if (LOGGER.isLoggable(Level.FINEST)) { LOGGER.log(Level.FINEST, "{0}= {1}", new Object[]{tmpNameC, tmpValueC}); } addParameter(tmpNameC.toString(), tmpValueC.toString()); } catch (Exception e) { decodeFailCount++; if (LOGGER.isLoggable(Level.FINEST)) { LOGGER.log(Level.FINEST, LogMessages.FINE_GRIZZLY_HTTP_PARAMETERS_DECODE_FAIL_DEBUG( origName.toString(), origValue.toString())); } else if (LOGGER.isLoggable( Level.INFO) && decodeFailCount == 1) { final String name = ((tmpNameC.getLength() > 0) ? tmpNameC.toString() : "unavailable"); final String value = ((tmpValueC.getLength() > 0) ? tmpValueC.toString() : "unavailable"); LOGGER.log(Level.INFO, LogMessages.INFO_GRIZZLY_HTTP_PARAMETERS_DECODE_FAIL_INFO( e.getMessage(), name, value)); LOGGER.log(Level.FINE, "Decoding stacktrace.", e); } } finally { tmpNameC.recycle(); tmpValueC.recycle(); } } while (pos < end); if (!LOGGER.isLoggable(Level.FINEST) && decodeFailCount > 1) { LOGGER.info( LogMessages.INFO_GRIZZLY_HTTP_PARAMETERS_MULTIPLE_DECODING_FAIL( decodeFailCount)); } } public void processParameters(final DataChunk data) { processParameters(data, encoding); } public void processParameters(final DataChunk data, final Charset encoding) { if (data == null || data.isNull() || data.getLength() <= 0) { return; } try { if (data.getType() == DataChunk.Type.Buffer) { final BufferChunk bc = data.getBufferChunk(); processParameters(bc.getBuffer(), bc.getStart(), bc.getLength(), encoding); } else { if (data.getType() != DataChunk.Type.Chars) { data.toChars(encoding); } final CharChunk cc = data.getCharChunk(); processParameters(cc.getChars(), cc.getStart(), cc.getLength()); } } catch (CharConversionException e) { throw new IllegalStateException(e); } }
Debug purpose
/** * Debug purpose */
public String paramsAsString() { StringBuilder sb = new StringBuilder(); for (final String s : paramHashValues.keySet()) { // END PWC 6057385 sb.append(s).append('='); ArrayList<String> v = paramHashValues.get(s); for (int i = 0, len = v.size(); i < len; i++) { sb.append(v.get(i)).append(','); } sb.append('\n'); } return sb.toString(); } // -------------------- Old code, needs rewrite -------------------- public void processParameters(String str) { int end = str.length(); int pos = 0; int decodeFailCount = 0; if (LOGGER.isLoggable(Level.FINEST)) { LOGGER.log(Level.FINEST, "Process parameters. String: {0}", str); } do { if (limit > -1 && parameterCount >= limit) { LOGGER.warning( LogMessages.WARNING_GRIZZLY_HTTP_SEVERE_GRIZZLY_HTTP_PARAMETERS_MAX_COUNT_FAIL( limit)); break; } boolean noEq = false; int valStart = -1; int valEnd = -1; int nameStart = pos; int nameEnd = str.indexOf('=', nameStart); int nameEnd2 = str.indexOf('&', nameStart); if (nameEnd2 == -1) { nameEnd2 = end; } if ((nameEnd2 != -1) && (nameEnd == -1 || nameEnd > nameEnd2)) { nameEnd = nameEnd2; noEq = true; valStart = nameEnd; valEnd = nameEnd; if (LOGGER.isLoggable(Level.FINEST)) { LOGGER.log(Level.FINEST, "no equal {0} {1} {2}", new Object[]{ nameStart, nameEnd, str.substring(nameStart, nameEnd) }); } } if (nameEnd == -1) { nameEnd = end; } if (!noEq) { valStart = nameEnd + 1; valEnd = str.indexOf('&', valStart); if (valEnd == -1) { valEnd = (valStart < end) ? end : valStart; } } pos = valEnd + 1; if (nameEnd <= nameStart) { continue; } if (LOGGER.isLoggable(Level.FINEST)) { LOGGER.log(Level.FINEST, "XXX {0} {1} {2} {3}", new Object[]{nameStart, nameEnd, valStart, valEnd}); } try { tmpNameC.append(str, nameStart, nameEnd - nameStart); tmpValueC.append(str, valStart, valEnd - valStart); if (LOGGER.isLoggable(Level.FINEST)) { LOGGER.log(Level.FINEST, "{0}= {1}", new Object[]{tmpNameC, tmpValueC}); } URLDecoder.decode(tmpNameC, true); URLDecoder.decode(tmpValueC, true); if (LOGGER.isLoggable(Level.FINEST)) { LOGGER.log(Level.FINEST, "{0}= {1}", new Object[]{tmpNameC, tmpValueC}); } addParameter(tmpNameC.toString(), tmpValueC.toString()); } catch (Exception e) { decodeFailCount++; if (LOGGER.isLoggable(Level.FINEST)) { LOGGER.log(Level.FINEST, LogMessages.FINE_GRIZZLY_HTTP_PARAMETERS_DECODE_FAIL_DEBUG( origName.toString(), origValue.toString())); } else if (LOGGER.isLoggable( Level.INFO) && decodeFailCount == 1) { final String name = ((tmpNameC.getLength() > 0) ? tmpNameC.toString() : "unavailable"); final String value = ((tmpValueC.getLength() > 0) ? tmpValueC.toString() : "unavailable"); LOGGER.log(Level.INFO, LogMessages.INFO_GRIZZLY_HTTP_PARAMETERS_DECODE_FAIL_INFO( e.getMessage(), name, value)); LOGGER.log(Level.FINE, "Decoding stacktrace.", e); } } finally { tmpNameC.recycle(); tmpValueC.recycle(); } } while (pos < end); if (!LOGGER.isLoggable(Level.FINEST) && decodeFailCount > 1) { LOGGER.info( LogMessages.INFO_GRIZZLY_HTTP_PARAMETERS_MULTIPLE_DECODING_FAIL( decodeFailCount)); } } }