| /* |
| * Copyright (c) 2014, 2015, Oracle and/or its affiliates. All rights reserved. |
| * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
| * |
| * This code is free software; you can redistribute it and/or modify it |
| * under the terms of the GNU General Public License version 2 only, as |
| * published by the Free Software Foundation. |
| * |
| * This code is distributed in the hope that it will be useful, but WITHOUT |
| * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
| * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
| * version 2 for more details (a copy is included in the LICENSE file that |
| * accompanied this code). |
| * |
| * You should have received a copy of the GNU General Public License version |
| * 2 along with this work; if not, write to the Free Software Foundation, |
| * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
| * |
| * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
| * or visit www.oracle.com if you need additional information or have any |
| * questions. |
| */ |
| package org.graalvm.compiler.lir.alloc.lsra; |
| |
| import static org.graalvm.compiler.lir.LIRValueUtil.isStackSlotValue; |
| import static jdk.vm.ci.code.ValueUtil.asRegister; |
| import static jdk.vm.ci.code.ValueUtil.isRegister; |
| |
| import org.graalvm.compiler.core.common.cfg.AbstractBlockBase; |
| import org.graalvm.compiler.debug.Debug; |
| import org.graalvm.compiler.debug.Debug.Scope; |
| import org.graalvm.compiler.debug.Indent; |
| import org.graalvm.compiler.lir.alloc.lsra.Interval.RegisterBinding; |
| import org.graalvm.compiler.lir.alloc.lsra.Interval.RegisterBindingLists; |
| import org.graalvm.compiler.lir.alloc.lsra.Interval.RegisterPriority; |
| import org.graalvm.compiler.lir.alloc.lsra.Interval.State; |
| import org.graalvm.compiler.options.Option; |
| import org.graalvm.compiler.options.OptionType; |
| import org.graalvm.compiler.options.OptionValue; |
| |
| import jdk.vm.ci.code.Register; |
| import jdk.vm.ci.meta.AllocatableValue; |
| |
| public class OptimizingLinearScanWalker extends LinearScanWalker { |
| |
| public static class Options { |
| // @formatter:off |
| @Option(help = "Enable LSRA optimization", type = OptionType.Debug) |
| public static final OptionValue<Boolean> LSRAOptimization = new OptionValue<>(false); |
| @Option(help = "LSRA optimization: Only split but do not reassign", type = OptionType.Debug) |
| public static final OptionValue<Boolean> LSRAOptSplitOnly = new OptionValue<>(false); |
| // @formatter:on |
| } |
| |
| OptimizingLinearScanWalker(LinearScan allocator, Interval unhandledFixedFirst, Interval unhandledAnyFirst) { |
| super(allocator, unhandledFixedFirst, unhandledAnyFirst); |
| } |
| |
| @SuppressWarnings("try") |
| @Override |
| protected void handleSpillSlot(Interval interval) { |
| assert interval.location() != null : "interval not assigned " + interval; |
| if (interval.canMaterialize()) { |
| assert !isStackSlotValue(interval.location()) : "interval can materialize but assigned to a stack slot " + interval; |
| return; |
| } |
| assert isStackSlotValue(interval.location()) : "interval not assigned to a stack slot " + interval; |
| try (Scope s1 = Debug.scope("LSRAOptimization")) { |
| Debug.log("adding stack to unhandled list %s", interval); |
| unhandledLists.addToListSortedByStartAndUsePositions(RegisterBinding.Stack, interval); |
| } |
| } |
| |
| @SuppressWarnings("unused") |
| private static void printRegisterBindingList(RegisterBindingLists list, RegisterBinding binding) { |
| for (Interval interval = list.get(binding); interval != Interval.EndMarker; interval = interval.next) { |
| Debug.log("%s", interval); |
| } |
| } |
| |
| @SuppressWarnings("try") |
| @Override |
| void walk() { |
| try (Scope s = Debug.scope("OptimizingLinearScanWalker")) { |
| for (AbstractBlockBase<?> block : allocator.sortedBlocks()) { |
| optimizeBlock(block); |
| } |
| } |
| super.walk(); |
| } |
| |
| @SuppressWarnings("try") |
| private void optimizeBlock(AbstractBlockBase<?> block) { |
| if (block.getPredecessorCount() == 1) { |
| int nextBlock = allocator.getFirstLirInstructionId(block); |
| try (Scope s1 = Debug.scope("LSRAOptimization")) { |
| Debug.log("next block: %s (%d)", block, nextBlock); |
| } |
| try (Indent indent0 = Debug.indent()) { |
| walkTo(nextBlock); |
| |
| try (Scope s1 = Debug.scope("LSRAOptimization")) { |
| boolean changed = true; |
| // we need to do this because the active lists might change |
| loop: while (changed) { |
| changed = false; |
| try (Indent indent1 = Debug.logAndIndent("Active intervals: (block %s [%d])", block, nextBlock)) { |
| for (Interval active = activeLists.get(RegisterBinding.Any); active != Interval.EndMarker; active = active.next) { |
| Debug.log("active (any): %s", active); |
| if (optimize(nextBlock, block, active, RegisterBinding.Any)) { |
| changed = true; |
| break loop; |
| } |
| } |
| for (Interval active = activeLists.get(RegisterBinding.Stack); active != Interval.EndMarker; active = active.next) { |
| Debug.log("active (stack): %s", active); |
| if (optimize(nextBlock, block, active, RegisterBinding.Stack)) { |
| changed = true; |
| break loop; |
| } |
| } |
| } |
| } |
| } |
| } |
| } |
| } |
| |
| @SuppressWarnings("try") |
| private boolean optimize(int currentPos, AbstractBlockBase<?> currentBlock, Interval currentInterval, RegisterBinding binding) { |
| // BEGIN initialize and sanity checks |
| assert currentBlock != null : "block must not be null"; |
| assert currentInterval != null : "interval must not be null"; |
| |
| assert currentBlock.getPredecessorCount() == 1 : "more than one predecessors -> optimization not possible"; |
| |
| if (!currentInterval.isSplitChild()) { |
| // interval is not a split child -> no need for optimization |
| return false; |
| } |
| |
| if (currentInterval.from() == currentPos) { |
| // the interval starts at the current position so no need for splitting |
| return false; |
| } |
| |
| // get current location |
| AllocatableValue currentLocation = currentInterval.location(); |
| assert currentLocation != null : "active intervals must have a location assigned!"; |
| |
| // get predecessor stuff |
| AbstractBlockBase<?> predecessorBlock = currentBlock.getPredecessors()[0]; |
| int predEndId = allocator.getLastLirInstructionId(predecessorBlock); |
| Interval predecessorInterval = currentInterval.getIntervalCoveringOpId(predEndId); |
| assert predecessorInterval != null : "variable not live at the end of the only predecessor! " + predecessorBlock + " -> " + currentBlock + " interval: " + currentInterval; |
| AllocatableValue predecessorLocation = predecessorInterval.location(); |
| assert predecessorLocation != null : "handled intervals must have a location assigned!"; |
| |
| // END initialize and sanity checks |
| |
| if (currentLocation.equals(predecessorLocation)) { |
| // locations are already equal -> nothing to optimize |
| return false; |
| } |
| |
| if (!isStackSlotValue(predecessorLocation) && !isRegister(predecessorLocation)) { |
| assert predecessorInterval.canMaterialize(); |
| // value is materialized -> no need for optimization |
| return false; |
| } |
| |
| assert isStackSlotValue(currentLocation) || isRegister(currentLocation) : "current location not a register or stack slot " + currentLocation; |
| |
| try (Indent indent = Debug.logAndIndent("location differs: %s vs. %s", predecessorLocation, currentLocation)) { |
| // split current interval at current position |
| Debug.log("splitting at position %d", currentPos); |
| |
| assert allocator.isBlockBegin(currentPos) && ((currentPos & 1) == 0) : "split pos must be even when on block boundary"; |
| |
| Interval splitPart = currentInterval.split(currentPos, allocator); |
| activeLists.remove(binding, currentInterval); |
| |
| assert splitPart.from() >= currentPosition : "cannot append new interval before current walk position"; |
| |
| // the currentSplitChild is needed later when moves are inserted for reloading |
| assert splitPart.currentSplitChild() == currentInterval : "overwriting wrong currentSplitChild"; |
| splitPart.makeCurrentSplitChild(); |
| |
| if (Debug.isLogEnabled()) { |
| Debug.log("left interval : %s", currentInterval.logString(allocator)); |
| Debug.log("right interval : %s", splitPart.logString(allocator)); |
| } |
| |
| if (Options.LSRAOptSplitOnly.getValue()) { |
| // just add the split interval to the unhandled list |
| unhandledLists.addToListSortedByStartAndUsePositions(RegisterBinding.Any, splitPart); |
| } else { |
| if (isRegister(predecessorLocation)) { |
| splitRegisterInterval(splitPart, asRegister(predecessorLocation)); |
| } else { |
| assert isStackSlotValue(predecessorLocation); |
| Debug.log("assigning interval %s to %s", splitPart, predecessorLocation); |
| splitPart.assignLocation(predecessorLocation); |
| // activate interval |
| activeLists.addToListSortedByCurrentFromPositions(RegisterBinding.Stack, splitPart); |
| splitPart.state = State.Active; |
| |
| splitStackInterval(splitPart); |
| } |
| } |
| } |
| return true; |
| } |
| |
| @SuppressWarnings("try") |
| private void splitRegisterInterval(Interval interval, Register reg) { |
| // collect current usage of registers |
| initVarsForAlloc(interval); |
| initUseLists(false); |
| spillExcludeActiveFixed(); |
| // spillBlockUnhandledFixed(cur); |
| assert unhandledLists.get(RegisterBinding.Fixed) == Interval.EndMarker : "must not have unhandled fixed intervals because all fixed intervals have a use at position 0"; |
| spillBlockInactiveFixed(interval); |
| spillCollectActiveAny(RegisterPriority.LiveAtLoopEnd); |
| spillCollectInactiveAny(interval); |
| |
| if (Debug.isLogEnabled()) { |
| try (Indent indent2 = Debug.logAndIndent("state of registers:")) { |
| for (Register register : availableRegs) { |
| int i = register.number; |
| try (Indent indent3 = Debug.logAndIndent("reg %d: usePos: %d, blockPos: %d, intervals: ", i, usePos[i], blockPos[i])) { |
| for (int j = 0; j < spillIntervals[i].size(); j++) { |
| Debug.log("%d ", spillIntervals[i].get(j).operandNumber); |
| } |
| } |
| } |
| } |
| } |
| |
| // the register must be free at least until this position |
| boolean needSplit = blockPos[reg.number] <= interval.to(); |
| |
| int splitPos = blockPos[reg.number]; |
| |
| assert splitPos > 0 : "invalid splitPos"; |
| assert needSplit || splitPos > interval.from() : "splitting interval at from"; |
| |
| Debug.log("assigning interval %s to %s", interval, reg); |
| interval.assignLocation(reg.asValue(interval.kind())); |
| if (needSplit) { |
| // register not available for full interval : so split it |
| splitWhenPartialRegisterAvailable(interval, splitPos); |
| } |
| |
| // perform splitting and spilling for all affected intervals |
| splitAndSpillIntersectingIntervals(reg); |
| |
| // activate interval |
| activeLists.addToListSortedByCurrentFromPositions(RegisterBinding.Any, interval); |
| interval.state = State.Active; |
| |
| } |
| } |