Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 20 of 182 for Postorder (0.14 sec)

  1. android/guava/src/com/google/common/graph/Traverser.java

        ImmutableSet<N> validated = validate(startNodes);
        return new Iterable<N>() {
          @Override
          public Iterator<N> iterator() {
            return newTraversal().postOrder(validated.iterator());
          }
        };
      }
    
      abstract Traversal<N> newTraversal();
    
      @SuppressWarnings("CheckReturnValue")
      private ImmutableSet<N> validate(Iterable<? extends N> startNodes) {
    Registered: Wed Jun 12 16:38:11 UTC 2024
    - Last Modified: Tue May 30 20:12:45 UTC 2023
    - 19.8K bytes
    - Viewed (0)
  2. guava/src/com/google/common/graph/Traverser.java

        ImmutableSet<N> validated = validate(startNodes);
        return new Iterable<N>() {
          @Override
          public Iterator<N> iterator() {
            return newTraversal().postOrder(validated.iterator());
          }
        };
      }
    
      abstract Traversal<N> newTraversal();
    
      @SuppressWarnings("CheckReturnValue")
      private ImmutableSet<N> validate(Iterable<? extends N> startNodes) {
    Registered: Wed Jun 12 16:38:11 UTC 2024
    - Last Modified: Tue May 30 20:12:45 UTC 2023
    - 19.8K bytes
    - Viewed (0)
  3. src/cmd/compile/internal/ssa/loopreschedchecks.go

    		return
    	}
    
    	backedges := backedges(f)
    	if len(backedges) == 0 { // no backedges means no rescheduling checks.
    		return
    	}
    
    	lastMems := findLastMems(f)
    
    	idom := f.Idom()
    	po := f.postorder()
    	// The ordering in the dominator tree matters; it's important that
    	// the walk of the dominator tree also be a preorder (i.e., a node is
    	// visited only after all its non-backedge predecessors have been visited).
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Tue Aug 22 21:17:10 UTC 2023
    - 16K bytes
    - Viewed (0)
  4. src/cmd/compile/internal/ssa/stackalloc.go

    	// Instead of iterating over f.Blocks, iterate over their postordering.
    	// Liveness information flows backward, so starting at the end
    	// increases the probability that we will stabilize quickly.
    	po := s.f.postorder()
    	for {
    		changed := false
    		for _, b := range po {
    			// Start with known live values at the end of the block
    			live.clear()
    			live.addAll(s.live[b.ID])
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu Feb 29 21:29:41 UTC 2024
    - 12.6K bytes
    - Viewed (0)
  5. src/cmd/compile/internal/ssa/branchelim.go

    			case OpLoad, OpAtomicLoad8, OpAtomicLoad32, OpAtomicLoad64, OpAtomicLoadPtr, OpAtomicLoadAcq32, OpAtomicLoadAcq64:
    				loadAddr.add(v.Args[0].ID)
    			case OpMove:
    				loadAddr.add(v.Args[1].ID)
    			}
    		}
    	}
    	po := f.postorder()
    	for {
    		n := loadAddr.size()
    		for _, b := range po {
    			for i := len(b.Values) - 1; i >= 0; i-- {
    				v := b.Values[i]
    				if !loadAddr.contains(v.ID) {
    					continue
    				}
    				for _, a := range v.Args {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed Nov 30 17:46:51 UTC 2022
    - 12.7K bytes
    - Viewed (0)
  6. src/cmd/compile/internal/ssa/check.go

    				}
    			}
    		}
    	}
    
    	// Check loop construction
    	if f.RegAlloc == nil && f.pass != nil { // non-nil pass allows better-targeted debug printing
    		ln := f.loopnest()
    		if !ln.hasIrreducible {
    			po := f.postorder() // use po to avoid unreachable blocks.
    			for _, b := range po {
    				for _, s := range b.Succs {
    					bb := s.Block()
    					if ln.b2l[b.ID] == nil && ln.b2l[bb.ID] != nil && bb != ln.b2l[bb.ID].header {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Tue Apr 09 16:41:23 UTC 2024
    - 17.6K bytes
    - Viewed (0)
  7. src/cmd/compile/internal/liveness/plive.go

    	// frees within the loop.
    	nvars := int32(len(lv.vars))
    	newlivein := bitvec.New(nvars)
    	newliveout := bitvec.New(nvars)
    
    	// Walk blocks in postorder ordering. This improves convergence.
    	po := lv.f.Postorder()
    
    	// Iterate through the blocks in reverse round-robin fashion. A work
    	// queue might be slightly faster. As is, the number of iterations is
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri Jun 07 15:22:22 UTC 2024
    - 45.2K bytes
    - Viewed (0)
  8. src/cmd/link/internal/ld/lib.go

    		if err != io.EOF {
    			log.Fatalf("reading input: %v", err)
    		}
    		return -1
    	}
    	return int(c)
    }
    
    type markKind uint8 // for postorder traversal
    const (
    	_ markKind = iota
    	visiting
    	visited
    )
    
    func postorder(libs []*sym.Library) []*sym.Library {
    	order := make([]*sym.Library, 0, len(libs)) // hold the result
    	mark := make(map[*sym.Library]markKind, len(libs))
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Tue May 21 18:45:27 UTC 2024
    - 88.6K bytes
    - Viewed (0)
  9. src/cmd/compile/internal/ssa/debug.go

    func (state *debugState) liveness() []*BlockDebug {
    	blockLocs := make([]*BlockDebug, state.f.NumBlocks())
    	counterTime := int32(1)
    
    	// Reverse postorder: visit a block after as many as possible of its
    	// predecessors have been visited.
    	po := state.f.Postorder()
    	converged := false
    
    	// The iteration rule is that by default, run until converged, but
    	// if a particular iteration count is specified, run that many
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Jun 10 19:44:43 UTC 2024
    - 58.4K bytes
    - Viewed (0)
  10. src/cmd/go/internal/list/list.go

    					p.Imports[i] = new
    				}
    			}
    			clear(m)
    		}
    	}
    
    	if listJsonFields.needAny("Deps", "DepsErrors") {
    		all := pkgs
    		// Make sure we iterate through packages in a postorder traversal,
    		// which load.PackageList guarantees. If *listDeps, then all is
    		// already in PackageList order. Otherwise, calling load.PackageList
    		// provides the guarantee. In the case of an import cycle, the last package
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri Feb 16 16:56:39 UTC 2024
    - 33.3K bytes
    - Viewed (0)
Back to top