- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 46 for SetArg (0.73 sec)
-
src/cmd/compile/internal/ssagen/phi.go
// Only set arguments that have been resolved. // For very wide CFGs, this significantly speeds up phi resolution. // See golang.org/issue/8225. if w := values[v.AuxInt]; w.Op != ssa.OpUnknown { v.SetArg(i, w) } } } // Walk children in dominator tree. for c := s.tree[b.ID].firstChild; c != nil; c = s.tree[c.ID].sibling { stk = append(stk, stackEntry{b: c}) } } }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Nov 18 17:59:44 UTC 2022 - 15.2K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/value.go
} func (v *Value) AddArgs(a ...*Value) { if v.Args == nil { v.resetArgs() // use argstorage } v.Args = append(v.Args, a...) for _, x := range a { x.Uses++ } } func (v *Value) SetArg(i int, w *Value) { v.Args[i].Uses-- v.Args[i] = w w.Uses++ } func (v *Value) SetArgs1(a *Value) { v.resetArgs() v.AddArg(a) } func (v *Value) SetArgs2(a, b *Value) { v.resetArgs()
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 03 16:40:22 UTC 2024 - 16.7K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/regalloc.go
vi := &s.values[a.ID] if vi.regs == 0 && !vi.rematerializeable { // Use the spill location. // This forces later liveness analysis to make the // value live at this point. v.SetArg(0, s.makeSpill(a, b)) } else if _, ok := a.Aux.(*ir.Name); ok && vi.rematerializeable { // Rematerializeable value with a gc.Node. This is the address of // a stack object (e.g. an LEAQ). Keep the object live.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Nov 21 17:49:56 UTC 2023 - 87.2K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/rewrite.go
// of a value so that a x.Uses==1 rule condition // fires reliably. for i, a := range v.Args { if a.Op != OpCopy { continue } aa := copySource(a) v.SetArg(i, aa) // If a, a copy, has a line boundary indicator, attempt to find a new value // to hold it. The first candidate is the value that will replace a (aa), // if it shares the same block and line and is eligible.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 19:02:52 UTC 2024 - 64.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/post_quantize.cc
if (auto qtype = QuantizedType::getQuantizedElementType(q.getArg().getType())) { rewriter.setInsertionPoint(op); rewriter.replaceOpWithNewOp<quantfork::DequantizeCastOp>( op, op.getResult().getType(), q.getArg()); return success(); } op.replaceAllUsesWith(q.getArg()); return success(); } return failure(); } };
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 05 07:39:40 UTC 2024 - 5.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/passes/post_quantize.cc
if (const QuantizedType qtype = QuantizedType::getQuantizedElementType(q.getArg().getType())) { rewriter.setInsertionPoint(op); rewriter.replaceOpWithNewOp<quantfork::DequantizeCastOp>( op, op.getResult().getType(), q.getArg()); return success(); } op.replaceAllUsesWith(q.getArg()); return success(); } return failure(); } };
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 05 07:39:40 UTC 2024 - 6.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/quantization/ir/ConvertConst.cc
LogicalResult QuantizedConstRewrite::matchAndRewrite( QuantizeCastOp qbarrier, PatternRewriter &rewriter) const { Attribute value; // Is the operand a constant? if (!matchPattern(qbarrier.getArg(), m_Constant(&value))) { return failure(); } // Does the qbarrier convert to a quantized type. This will not be true // if a quantized type has not yet been chosen or if the cast to an equivalent
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 4.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfr/ir/tfr_ops.cc
if (preceding_cast.getArg().getDefiningOp()) { auto redundant_cast = preceding_cast.getArg().getDefiningOp<CastOp>(); if (!redundant_cast || redundant_cast.getArg().getType() != preceding_cast.getOut().getType()) { return failure(); } raw_data_op.getOutput().replaceAllUsesWith(redundant_cast.getArg()); } else {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Nov 21 16:55:41 UTC 2023 - 38.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/common/ir/QuantOps.cc
// Matches x -> [scast -> scast] -> y, replacing the second scast with the // value of x if the casts invert each other. auto srcScastOp = getArg().getDefiningOp<StorageCastOp>(); if (!srcScastOp || srcScastOp.getArg().getType() != getType()) return OpFoldResult(); return srcScastOp.getArg(); } /// The quantization specification should match the expressed type.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 5.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/quantization/ir/QuantOps.cc
// Matches x -> [scast -> scast] -> y, replacing the second scast with the // value of x if the casts invert each other. auto srcScastOp = getArg().getDefiningOp<StorageCastOp>(); if (!srcScastOp || srcScastOp.getArg().getType() != getType()) return OpFoldResult(); return srcScastOp.getArg(); } /// The quantization specification should match the expressed type.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 5.3K bytes - Viewed (0)