31#include "mlir/Pass/Pass.h"
32#include "mlir/Transforms/DialectConversion.h"
33#include "llvm/ADT/APInt.h"
34#include "llvm/ADT/PointerUnion.h"
35#include "llvm/Support/Debug.h"
38#define DEBUG_TYPE "comb-to-synth"
41#define GEN_PASS_DEF_CONVERTCOMBTOSYNTH
42#include "circt/Conversion/Passes.h.inc"
53static SmallVector<Value>
extractBits(OpBuilder &builder, Value val) {
54 SmallVector<Value> bits;
55 comb::extractBits(builder, val, bits);
66template <
bool isLeftShift>
68 Value shiftAmount, int64_t maxShiftAmount,
69 llvm::function_ref<Value(int64_t)> getPadding,
70 llvm::function_ref<Value(int64_t)> getExtract) {
75 SmallVector<Value> nodes;
76 nodes.reserve(maxShiftAmount);
77 for (int64_t i = 0; i < maxShiftAmount; ++i) {
78 Value extract = getExtract(i);
79 Value padding = getPadding(i);
82 nodes.push_back(extract);
96 auto outOfBoundsValue = getPadding(maxShiftAmount);
97 assert(outOfBoundsValue &&
"outOfBoundsValue must be valid");
101 comb::constructMuxTree(rewriter, loc, bits, nodes, outOfBoundsValue);
104 auto inBound = rewriter.createOrFold<comb::ICmpOp>(
105 loc, ICmpPredicate::ult, shiftAmount,
109 return rewriter.createOrFold<
comb::MuxOp>(loc, inBound, result,
116 Value b, Value carry) {
118 auto aXnorB = comb::XorOp::create(rewriter, loc, ValueRange{a, b},
true);
120 comb::AndOp::create(rewriter, loc, ValueRange{carry, aXnorB},
true);
121 auto aAndB = comb::AndOp::create(rewriter, loc, ValueRange{a, b},
true);
122 return comb::OrOp::create(rewriter, loc, ValueRange{andOp, aAndB},
true);
127 val.getLoc(), val, val.getType().getIntOrFloatBitWidth() - 1, 1);
132 val.getLoc(), val, 0, val.getType().getIntOrFloatBitWidth() - 1);
137using ConstantOrValue = llvm::PointerUnion<Value, mlir::IntegerAttr>;
142 Value value, llvm::SmallVectorImpl<ConstantOrValue> &values) {
144 if (value.getType().isInteger(0))
149 int64_t totalUnknownBits = 0;
150 for (
auto concatInput : llvm::reverse(concat.getInputs())) {
155 totalUnknownBits += unknownBits;
157 return totalUnknownBits;
162 values.push_back(constant.getValueAttr());
168 values.push_back(value);
169 return hw::getBitWidth(value.getType());
175 llvm::SmallVectorImpl<ConstantOrValue> &constantOrValues,
177 uint32_t bitPos = 0, unknownPos = 0;
178 APInt result(width, 0);
179 for (
auto constantOrValue : constantOrValues) {
181 if (
auto constant = dyn_cast<IntegerAttr>(constantOrValue)) {
182 elemWidth = constant.getValue().getBitWidth();
183 result.insertBits(constant.getValue(), bitPos);
185 elemWidth = hw::getBitWidth(cast<Value>(constantOrValue).getType());
186 assert(elemWidth >= 0 &&
"unknown bit width");
187 assert(elemWidth + unknownPos < 32 &&
"unknown bit width too large");
189 uint32_t usedBits = (mask >> unknownPos) & ((1 << elemWidth) - 1);
190 result.insertBits(APInt(elemWidth, usedBits), bitPos);
191 unknownPos += elemWidth;
203 ConversionPatternRewriter &rewriter, int64_t maxEmulationUnknownBits,
205 llvm::function_ref<APInt(
const APInt &,
const APInt &)> emulate) {
206 SmallVector<ConstantOrValue> lhsValues, rhsValues;
208 assert(op->getNumResults() == 1 && op->getNumOperands() == 2 &&
209 "op must be a single result binary operation");
211 auto lhs = op->getOperand(0);
212 auto rhs = op->getOperand(1);
213 auto width = op->getResult(0).getType().getIntOrFloatBitWidth();
214 auto loc = op->getLoc();
219 if (numLhsUnknownBits < 0 || numRhsUnknownBits < 0)
222 int64_t totalUnknownBits = numLhsUnknownBits + numRhsUnknownBits;
223 if (totalUnknownBits > maxEmulationUnknownBits)
226 SmallVector<Value> emulatedResults;
227 emulatedResults.reserve(1 << totalUnknownBits);
230 DenseMap<IntegerAttr, hw::ConstantOp> constantPool;
232 auto attr = rewriter.getIntegerAttr(rewriter.getIntegerType(width), value);
233 auto it = constantPool.find(attr);
234 if (it != constantPool.end())
237 constantPool[attr] = constant;
241 for (uint32_t lhsMask = 0, lhsMaskEnd = 1 << numLhsUnknownBits;
242 lhsMask < lhsMaskEnd; ++lhsMask) {
244 for (uint32_t rhsMask = 0, rhsMaskEnd = 1 << numRhsUnknownBits;
245 rhsMask < rhsMaskEnd; ++rhsMask) {
248 emulatedResults.push_back(
getConstant(emulate(lhsValue, rhsValue)));
253 SmallVector<Value> selectors;
254 selectors.reserve(totalUnknownBits);
255 for (
auto &concatedValues : {rhsValues, lhsValues})
256 for (
auto valueOrConstant : concatedValues) {
257 auto value = dyn_cast<Value>(valueOrConstant);
263 assert(totalUnknownBits ==
static_cast<int64_t
>(selectors.size()) &&
264 "number of selectors must match");
265 auto muxed = constructMuxTree(rewriter, loc, selectors, emulatedResults,
283 matchAndRewrite(
AndOp op, OpAdaptor adaptor,
284 ConversionPatternRewriter &rewriter)
const override {
285 SmallVector<bool> nonInverts(adaptor.getInputs().size(),
false);
286 replaceOpWithNewOpAndCopyNamehint<synth::aig::AndInverterOp>(
287 rewriter, op, adaptor.getInputs(), nonInverts);
297 matchAndRewrite(
OrOp op, OpAdaptor adaptor,
298 ConversionPatternRewriter &rewriter)
const override {
300 SmallVector<bool> allInverts(adaptor.getInputs().size(),
true);
301 auto andOp = synth::aig::AndInverterOp::create(
302 rewriter, op.getLoc(), adaptor.getInputs(), allInverts);
303 replaceOpWithNewOpAndCopyNamehint<synth::aig::AndInverterOp>(
314 matchAndRewrite(
XorOp op, OpAdaptor adaptor,
315 ConversionPatternRewriter &rewriter)
const override {
316 SmallVector<bool> inverted(adaptor.getInputs().size(),
false);
317 replaceOpWithNewOpAndCopyNamehint<synth::XorInverterOp>(
318 rewriter, op, adaptor.getInputs(), inverted);
324struct SynthXorInverterOpConversion
329 matchAndRewrite(synth::XorInverterOp op, OpAdaptor adaptor,
330 ConversionPatternRewriter &rewriter)
const override {
331 if (op.getNumOperands() != 2)
337 auto inputs = adaptor.getInputs();
338 auto allNotInverts = op.getInverted();
339 std::array<bool, 2> allInverts = {!allNotInverts[0], !allNotInverts[1]};
341 auto notAAndNotB = synth::aig::AndInverterOp::create(rewriter, op.getLoc(),
343 auto aAndB = synth::aig::AndInverterOp::create(rewriter, op.getLoc(),
344 inputs, allNotInverts);
346 replaceOpWithNewOpAndCopyNamehint<synth::aig::AndInverterOp>(
347 rewriter, op, notAAndNotB, aAndB,
354template <
typename OpTy>
359 matchAndRewrite(OpTy op, OpAdaptor adaptor,
360 ConversionPatternRewriter &rewriter)
const override {
367 ConversionPatternRewriter &rewriter) {
369 switch (operands.size()) {
371 llvm_unreachable(
"cannot be called with empty operand range");
378 return OpTy::create(rewriter, op.getLoc(), ValueRange{lhs, rhs},
true);
380 auto firstHalf = operands.size() / 2;
385 return OpTy::create(rewriter, op.getLoc(), ValueRange{lhs, rhs},
true);
395 matchAndRewrite(
MuxOp op, OpAdaptor adaptor,
396 ConversionPatternRewriter &rewriter)
const override {
397 Value cond = op.getCond();
398 auto trueVal = op.getTrueValue();
399 auto falseVal = op.getFalseValue();
401 if (!op.getType().isInteger()) {
403 auto widthType = rewriter.getIntegerType(hw::getBitWidth(op.getType()));
411 if (!trueVal.getType().isInteger(1))
412 cond = comb::ReplicateOp::create(rewriter, op.getLoc(), trueVal.getType(),
417 synth::aig::AndInverterOp::create(rewriter, op.getLoc(), cond, trueVal);
418 auto rhs = synth::aig::AndInverterOp::create(rewriter, op.getLoc(), cond,
419 falseVal,
true,
false);
421 Value result = comb::OrOp::create(rewriter, op.getLoc(), lhs, rhs);
423 if (result.getType() != op.getType())
435enum AdderArchitecture { RippleCarry, Sklanskey, KoggeStone, BrentKung };
436AdderArchitecture determineAdderArch(Operation *op, int64_t width) {
437 auto strAttr = op->getAttrOfType<StringAttr>(
"synth.test.arch");
439 return llvm::StringSwitch<AdderArchitecture>(strAttr.getValue())
440 .Case(
"SKLANSKEY", Sklanskey)
441 .Case(
"KOGGE-STONE", KoggeStone)
442 .Case(
"BRENT-KUNG", BrentKung)
443 .Case(
"RIPPLE-CARRY", RippleCarry);
453 return AdderArchitecture::RippleCarry;
458 return AdderArchitecture::Sklanskey;
462 return AdderArchitecture::KoggeStone;
472void lowerKoggeStonePrefixTree(OpBuilder &builder, Location loc,
473 SmallVector<Value> &pPrefix,
474 SmallVector<Value> &gPrefix) {
476 auto width =
static_cast<int64_t
>(pPrefix.size());
477 assert(width ==
static_cast<int64_t
>(gPrefix.size()));
478 SmallVector<Value> pPrefixNew = pPrefix;
479 SmallVector<Value> gPrefixNew = gPrefix;
482 for (int64_t stride = 1; stride < width; stride *= 2) {
484 for (int64_t i = stride; i < width; ++i) {
485 int64_t j = i - stride;
488 Value andPG = comb::AndOp::create(builder, loc, pPrefix[i], gPrefix[j]);
489 gPrefixNew[i] = comb::OrOp::create(builder, loc, gPrefix[i], andPG);
492 pPrefixNew[i] = comb::AndOp::create(builder, loc, pPrefix[i], pPrefix[j]);
495 pPrefix = pPrefixNew;
496 gPrefix = gPrefixNew;
501 for (int64_t stride = 1; stride < width; stride *= 2) {
503 <<
"--------------------------------------- Kogge-Stone Stage "
505 for (int64_t i = stride; i < width; ++i) {
506 int64_t j = i - stride;
508 llvm::dbgs() <<
"G" << i << stage + 1 <<
" = G" << i << stage
509 <<
" OR (P" << i << stage <<
" AND G" << j << stage
513 llvm::dbgs() <<
"P" << i << stage + 1 <<
" = P" << i << stage
514 <<
" AND P" << j << stage <<
"\n";
523void lowerSklanskeyPrefixTree(OpBuilder &builder, Location loc,
524 SmallVector<Value> &pPrefix,
525 SmallVector<Value> &gPrefix) {
526 auto width =
static_cast<int64_t
>(pPrefix.size());
527 assert(width ==
static_cast<int64_t
>(gPrefix.size()));
528 SmallVector<Value> pPrefixNew = pPrefix;
529 SmallVector<Value> gPrefixNew = gPrefix;
530 for (int64_t stride = 1; stride < width; stride *= 2) {
531 for (int64_t i = stride; i < width; i += 2 * stride) {
532 for (int64_t k = 0; k < stride && i + k < width; ++k) {
538 comb::AndOp::create(builder, loc, pPrefix[idx], gPrefix[j]);
539 gPrefixNew[idx] = comb::OrOp::create(builder, loc, gPrefix[idx], andPG);
543 comb::AndOp::create(builder, loc, pPrefix[idx], pPrefix[j]);
547 pPrefix = pPrefixNew;
548 gPrefix = gPrefixNew;
553 for (int64_t stride = 1; stride < width; stride *= 2) {
554 llvm::dbgs() <<
"--------------------------------------- Sklanskey Stage "
556 for (int64_t i = stride; i < width; i += 2 * stride) {
557 for (int64_t k = 0; k < stride && i + k < width; ++k) {
561 llvm::dbgs() <<
"G" << idx << stage + 1 <<
" = G" << idx << stage
562 <<
" OR (P" << idx << stage <<
" AND G" << j << stage
566 llvm::dbgs() <<
"P" << idx << stage + 1 <<
" = P" << idx << stage
567 <<
" AND P" << j << stage <<
"\n";
578void lowerBrentKungPrefixTree(OpBuilder &builder, Location loc,
579 SmallVector<Value> &pPrefix,
580 SmallVector<Value> &gPrefix) {
581 auto width =
static_cast<int64_t
>(pPrefix.size());
582 assert(width ==
static_cast<int64_t
>(gPrefix.size()));
583 SmallVector<Value> pPrefixNew = pPrefix;
584 SmallVector<Value> gPrefixNew = gPrefix;
588 for (stride = 1; stride < width; stride *= 2) {
589 for (int64_t i = stride * 2 - 1; i < width; i += stride * 2) {
590 int64_t j = i - stride;
593 Value andPG = comb::AndOp::create(builder, loc, pPrefix[i], gPrefix[j]);
594 gPrefixNew[i] = comb::OrOp::create(builder, loc, gPrefix[i], andPG);
597 pPrefixNew[i] = comb::AndOp::create(builder, loc, pPrefix[i], pPrefix[j]);
599 pPrefix = pPrefixNew;
600 gPrefix = gPrefixNew;
604 for (; stride > 0; stride /= 2) {
605 for (int64_t i = stride * 3 - 1; i < width; i += stride * 2) {
606 int64_t j = i - stride;
609 Value andPG = comb::AndOp::create(builder, loc, pPrefix[i], gPrefix[j]);
610 gPrefixNew[i] = comb::OrOp::create(builder, loc, gPrefix[i], andPG);
613 pPrefixNew[i] = comb::AndOp::create(builder, loc, pPrefix[i], pPrefix[j]);
615 pPrefix = pPrefixNew;
616 gPrefix = gPrefixNew;
621 for (stride = 1; stride < width; stride *= 2) {
622 llvm::dbgs() <<
"--------------------------------------- Brent-Kung FW "
623 << stage <<
" : Stride " << stride <<
"\n";
624 for (int64_t i = stride * 2 - 1; i < width; i += stride * 2) {
625 int64_t j = i - stride;
628 llvm::dbgs() <<
"G" << i << stage + 1 <<
" = G" << i << stage
629 <<
" OR (P" << i << stage <<
" AND G" << j << stage
633 llvm::dbgs() <<
"P" << i << stage + 1 <<
" = P" << i << stage
634 <<
" AND P" << j << stage <<
"\n";
639 for (; stride > 0; stride /= 2) {
640 if (stride * 3 - 1 < width)
641 llvm::dbgs() <<
"--------------------------------------- Brent-Kung BW "
642 << stage <<
" : Stride " << stride <<
"\n";
644 for (int64_t i = stride * 3 - 1; i < width; i += stride * 2) {
645 int64_t j = i - stride;
648 llvm::dbgs() <<
"G" << i << stage + 1 <<
" = G" << i << stage
649 <<
" OR (P" << i << stage <<
" AND G" << j << stage
653 llvm::dbgs() <<
"P" << i << stage + 1 <<
" = P" << i << stage
654 <<
" AND P" << j << stage <<
"\n";
662class LazyKoggeStonePrefixTree {
664 LazyKoggeStonePrefixTree(OpBuilder &builder, Location loc, int64_t width,
665 ArrayRef<Value> pPrefix, ArrayRef<Value> gPrefix)
666 : builder(builder), loc(loc), width(width) {
667 assert(width > 0 &&
"width must be positive");
668 for (int64_t i = 0; i < width; ++i)
669 prefixCache[{0, i}] = {pPrefix[i], gPrefix[i]};
673 std::pair<Value, Value> getFinal(int64_t i) {
674 assert(i >= 0 && i < width &&
"i out of bounds");
676 return getGroupAndPropagate(llvm::Log2_64_Ceil(width), i);
684 std::pair<Value, Value> getGroupAndPropagate(int64_t level, int64_t i);
688 DenseMap<std::pair<int64_t, int64_t>, std::pair<Value, Value>> prefixCache;
691std::pair<Value, Value>
692LazyKoggeStonePrefixTree::getGroupAndPropagate(int64_t level, int64_t i) {
693 assert(i < width &&
"i out of bounds");
694 auto key = std::make_pair(level, i);
695 auto it = prefixCache.find(key);
696 if (it != prefixCache.end())
699 assert(level > 0 &&
"If the level is 0, we should have hit the cache");
701 int64_t previousStride = 1ULL << (level - 1);
702 if (i < previousStride) {
704 auto [propagateI, generateI] = getGroupAndPropagate(level - 1, i);
705 prefixCache[key] = {propagateI, generateI};
706 return prefixCache[key];
709 int64_t j = i - previousStride;
710 auto [propagateI, generateI] = getGroupAndPropagate(level - 1, i);
711 auto [propagateJ, generateJ] = getGroupAndPropagate(level - 1, j);
713 Value andPG = comb::AndOp::create(builder, loc, propagateI, generateJ);
714 Value newGenerate = comb::OrOp::create(builder, loc, generateI, andPG);
717 comb::AndOp::create(builder, loc, propagateI, propagateJ);
718 prefixCache[key] = {newPropagate, newGenerate};
719 return prefixCache[key];
726 matchAndRewrite(
AddOp op, OpAdaptor adaptor,
727 ConversionPatternRewriter &rewriter)
const override {
728 auto inputs = adaptor.getInputs();
731 if (inputs.size() != 2)
734 auto width = op.getType().getIntOrFloatBitWidth();
737 replaceOpWithNewOpAndCopyNamehint<hw::ConstantOp>(rewriter, op,
743 auto arch = determineAdderArch(op, width);
744 if (arch == AdderArchitecture::RippleCarry)
745 return lowerRippleCarryAdder(op, inputs, rewriter);
746 return lowerParallelPrefixAdder(op, inputs, rewriter);
751 lowerRippleCarryAdder(
comb::AddOp op, ValueRange inputs,
752 ConversionPatternRewriter &rewriter)
const {
753 auto width = op.getType().getIntOrFloatBitWidth();
759 SmallVector<Value> results;
760 results.resize(width);
761 for (int64_t i = 0; i < width; ++i) {
762 SmallVector<Value> xorOperands = {aBits[i], bBits[i]};
764 xorOperands.push_back(carry);
768 results[width - i - 1] =
769 comb::XorOp::create(rewriter, op.getLoc(), xorOperands,
true);
778 carry = comb::AndOp::create(rewriter, op.getLoc(),
779 ValueRange{aBits[i], bBits[i]},
true);
786 LLVM_DEBUG(llvm::dbgs() <<
"Lower comb.add to Ripple-Carry Adder of width "
789 replaceOpWithNewOpAndCopyNamehint<comb::ConcatOp>(rewriter, op, results);
797 lowerParallelPrefixAdder(
comb::AddOp op, ValueRange inputs,
798 ConversionPatternRewriter &rewriter)
const {
799 auto width = op.getType().getIntOrFloatBitWidth();
805 SmallVector<Value> p, g;
809 for (
auto [aBit, bBit] :
llvm::zip(aBits, bBits)) {
811 p.push_back(comb::XorOp::create(rewriter, op.getLoc(), aBit, bBit));
813 g.push_back(comb::AndOp::create(rewriter, op.getLoc(), aBit, bBit));
817 llvm::dbgs() <<
"Lower comb.add to Parallel-Prefix of width " << width
818 <<
"\n--------------------------------------- Init\n";
820 for (int64_t i = 0; i < width; ++i) {
822 llvm::dbgs() <<
"P0" << i <<
" = A" << i <<
" XOR B" << i <<
"\n";
824 llvm::dbgs() <<
"G0" << i <<
" = A" << i <<
" AND B" << i <<
"\n";
829 SmallVector<Value> pPrefix = p;
830 SmallVector<Value> gPrefix = g;
833 auto arch = determineAdderArch(op, width);
836 case AdderArchitecture::RippleCarry:
837 llvm_unreachable(
"Ripple-Carry should be handled separately");
839 case AdderArchitecture::Sklanskey:
840 lowerSklanskeyPrefixTree(rewriter, op.getLoc(), pPrefix, gPrefix);
842 case AdderArchitecture::KoggeStone:
843 lowerKoggeStonePrefixTree(rewriter, op.getLoc(), pPrefix, gPrefix);
845 case AdderArchitecture::BrentKung:
846 lowerBrentKungPrefixTree(rewriter, op.getLoc(), pPrefix, gPrefix);
852 SmallVector<Value> results;
853 results.resize(width);
855 results[width - 1] = p[0];
859 for (int64_t i = 1; i < width; ++i)
860 results[width - 1 - i] =
861 comb::XorOp::create(rewriter, op.getLoc(), p[i], gPrefix[i - 1]);
863 replaceOpWithNewOpAndCopyNamehint<comb::ConcatOp>(rewriter, op, results);
866 llvm::dbgs() <<
"--------------------------------------- Completion\n"
868 for (int64_t i = 1; i < width; ++i)
869 llvm::dbgs() <<
"RES" << i <<
" = P" << i <<
" XOR G" << i - 1 <<
"\n";
880 matchAndRewrite(
MulOp op, OpAdaptor adaptor,
881 ConversionPatternRewriter &rewriter)
const override {
882 if (adaptor.getInputs().size() != 2)
885 Location loc = op.getLoc();
886 Value
a = adaptor.getInputs()[0];
887 Value
b = adaptor.getInputs()[1];
888 unsigned width = op.getType().getIntOrFloatBitWidth();
897 SmallVector<Value> aBits =
extractBits(rewriter, a);
898 SmallVector<Value> bBits =
extractBits(rewriter, b);
903 SmallVector<SmallVector<Value>> partialProducts;
904 partialProducts.reserve(width);
905 for (
unsigned i = 0; i < width; ++i) {
906 SmallVector<Value> row(i, falseValue);
909 for (
unsigned j = 0; i + j < width; ++j)
911 rewriter.createOrFold<
comb::AndOp>(loc, aBits[j], bBits[i]));
913 partialProducts.push_back(row);
918 rewriter.replaceOp(op, partialProducts[0][0]);
924 auto addends = comp.compressToHeight(rewriter, 2);
927 auto newAdd = comb::AddOp::create(rewriter, loc, addends,
true);
933template <
typename OpTy>
935 DivModOpConversionBase(MLIRContext *
context, int64_t maxEmulationUnknownBits)
937 maxEmulationUnknownBits(maxEmulationUnknownBits) {
938 assert(maxEmulationUnknownBits < 32 &&
939 "maxEmulationUnknownBits must be less than 32");
941 const int64_t maxEmulationUnknownBits;
944struct CombDivUOpConversion : DivModOpConversionBase<DivUOp> {
945 using DivModOpConversionBase<
DivUOp>::DivModOpConversionBase;
947 matchAndRewrite(
DivUOp op, OpAdaptor adaptor,
948 ConversionPatternRewriter &rewriter)
const override {
950 if (llvm::succeeded(comb::convertDivUByPowerOfTwo(op, rewriter)))
956 rewriter, maxEmulationUnknownBits, op,
957 [](
const APInt &lhs,
const APInt &rhs) {
960 return APInt::getZero(rhs.getBitWidth());
961 return lhs.udiv(rhs);
966struct CombModUOpConversion : DivModOpConversionBase<ModUOp> {
967 using DivModOpConversionBase<
ModUOp>::DivModOpConversionBase;
969 matchAndRewrite(
ModUOp op, OpAdaptor adaptor,
970 ConversionPatternRewriter &rewriter)
const override {
972 if (llvm::succeeded(comb::convertModUByPowerOfTwo(op, rewriter)))
978 rewriter, maxEmulationUnknownBits, op,
979 [](
const APInt &lhs,
const APInt &rhs) {
982 return APInt::getZero(rhs.getBitWidth());
983 return lhs.urem(rhs);
988struct CombDivSOpConversion : DivModOpConversionBase<DivSOp> {
989 using DivModOpConversionBase<
DivSOp>::DivModOpConversionBase;
992 matchAndRewrite(
DivSOp op, OpAdaptor adaptor,
993 ConversionPatternRewriter &rewriter)
const override {
997 rewriter, maxEmulationUnknownBits, op,
998 [](
const APInt &lhs,
const APInt &rhs) {
1001 return APInt::getZero(rhs.getBitWidth());
1002 return lhs.sdiv(rhs);
1007struct CombModSOpConversion : DivModOpConversionBase<ModSOp> {
1008 using DivModOpConversionBase<
ModSOp>::DivModOpConversionBase;
1010 matchAndRewrite(
ModSOp op, OpAdaptor adaptor,
1011 ConversionPatternRewriter &rewriter)
const override {
1015 rewriter, maxEmulationUnknownBits, op,
1016 [](
const APInt &lhs,
const APInt &rhs) {
1019 return APInt::getZero(rhs.getBitWidth());
1020 return lhs.srem(rhs);
1029 static Value constructRippleCarry(Location loc, Value a, Value b,
1031 ConversionPatternRewriter &rewriter) {
1039 for (
auto [aBit, bBit] :
llvm::zip(aBits, bBits)) {
1041 rewriter.createOrFold<
comb::XorOp>(loc, aBit, bBit,
true);
1042 auto aEqualB = rewriter.createOrFold<synth::aig::AndInverterOp>(
1043 loc, aBitXorBBit,
true);
1044 auto pred = rewriter.createOrFold<synth::aig::AndInverterOp>(
1045 loc, aBit, bBit,
true,
false);
1047 auto aBitAndBBit = rewriter.createOrFold<
comb::AndOp>(
1048 loc, ValueRange{aEqualB,
acc},
true);
1049 acc = rewriter.createOrFold<
comb::OrOp>(loc, pred, aBitAndBBit,
true);
1062 static Value computePrefixComparison(ConversionPatternRewriter &rewriter,
1063 Location loc, SmallVector<Value> pPrefix,
1064 SmallVector<Value> gPrefix,
1065 bool includeEq, AdderArchitecture arch) {
1066 auto width = pPrefix.size();
1067 Value finalGroup, finalPropagate;
1070 case AdderArchitecture::RippleCarry:
1071 llvm_unreachable(
"Ripple-Carry should be handled separately");
1073 case AdderArchitecture::Sklanskey: {
1074 lowerSklanskeyPrefixTree(rewriter, loc, pPrefix, gPrefix);
1075 finalGroup = gPrefix[width - 1];
1076 finalPropagate = pPrefix[width - 1];
1079 case AdderArchitecture::KoggeStone:
1082 std::tie(finalPropagate, finalGroup) =
1083 LazyKoggeStonePrefixTree(rewriter, loc, width, pPrefix, gPrefix)
1084 .getFinal(width - 1);
1086 case AdderArchitecture::BrentKung: {
1087 lowerBrentKungPrefixTree(rewriter, loc, pPrefix, gPrefix);
1088 finalGroup = gPrefix[width - 1];
1089 finalPropagate = pPrefix[width - 1];
1098 return comb::OrOp::create(rewriter, loc, finalGroup, finalPropagate);
1107 static Value constructUnsignedCompare(Operation *op, Location loc, Value a,
1108 Value b,
bool isLess,
bool includeEq,
1109 ConversionPatternRewriter &rewriter) {
1113 auto width =
a.getType().getIntOrFloatBitWidth();
1116 auto arch = determineAdderArch(op, width);
1117 if (arch == AdderArchitecture::RippleCarry)
1118 return constructRippleCarry(loc, a, b, includeEq, rewriter);
1129 SmallVector<Value> eq, gt;
1136 for (
auto [aBit, bBit] :
llvm::zip(aBits, bBits)) {
1138 auto xorBit = comb::XorOp::create(rewriter, loc, aBit, bBit);
1139 eq.push_back(comb::XorOp::create(rewriter, loc, xorBit, one));
1142 auto notA = comb::XorOp::create(rewriter, loc, aBit, one);
1143 gt.push_back(comb::AndOp::create(rewriter, loc, notA, bBit));
1146 return computePrefixComparison(rewriter, loc, std::move(eq), std::move(gt),
1151 matchAndRewrite(ICmpOp op, OpAdaptor adaptor,
1152 ConversionPatternRewriter &rewriter)
const override {
1153 auto lhs = adaptor.getLhs();
1154 auto rhs = adaptor.getRhs();
1156 switch (op.getPredicate()) {
1160 case ICmpPredicate::eq:
1161 case ICmpPredicate::ceq: {
1163 auto xorOp = rewriter.createOrFold<
comb::XorOp>(op.getLoc(), lhs, rhs);
1165 SmallVector<bool> allInverts(xorBits.size(),
true);
1166 replaceOpWithNewOpAndCopyNamehint<synth::aig::AndInverterOp>(
1167 rewriter, op, xorBits, allInverts);
1171 case ICmpPredicate::ne:
1172 case ICmpPredicate::cne: {
1174 auto xorOp = rewriter.createOrFold<
comb::XorOp>(op.getLoc(), lhs, rhs);
1175 replaceOpWithNewOpAndCopyNamehint<comb::OrOp>(
1176 rewriter, op,
extractBits(rewriter, xorOp),
true);
1180 case ICmpPredicate::uge:
1181 case ICmpPredicate::ugt:
1182 case ICmpPredicate::ule:
1183 case ICmpPredicate::ult: {
1184 bool isLess = op.getPredicate() == ICmpPredicate::ult ||
1185 op.getPredicate() == ICmpPredicate::ule;
1186 bool includeEq = op.getPredicate() == ICmpPredicate::uge ||
1187 op.getPredicate() == ICmpPredicate::ule;
1189 constructUnsignedCompare(op, op.getLoc(), lhs,
1190 rhs, isLess, includeEq,
1194 case ICmpPredicate::slt:
1195 case ICmpPredicate::sle:
1196 case ICmpPredicate::sgt:
1197 case ICmpPredicate::sge: {
1198 if (lhs.getType().getIntOrFloatBitWidth() == 0)
1199 return rewriter.notifyMatchFailure(
1200 op.getLoc(),
"i0 signed comparison is unsupported");
1201 bool isLess = op.getPredicate() == ICmpPredicate::slt ||
1202 op.getPredicate() == ICmpPredicate::sle;
1203 bool includeEq = op.getPredicate() == ICmpPredicate::sge ||
1204 op.getPredicate() == ICmpPredicate::sle;
1213 auto sameSignResult = constructUnsignedCompare(
1214 op, op.getLoc(), aRest, bRest, isLess, includeEq, rewriter);
1218 comb::XorOp::create(rewriter, op.getLoc(), signA, signB);
1221 Value diffSignResult = isLess ? signA : signB;
1224 replaceOpWithNewOpAndCopyNamehint<comb::MuxOp>(
1225 rewriter, op, signsDiffer, diffSignResult, sameSignResult);
1236 matchAndRewrite(
ParityOp op, OpAdaptor adaptor,
1237 ConversionPatternRewriter &rewriter)
const override {
1239 replaceOpWithNewOpAndCopyNamehint<comb::XorOp>(
1240 rewriter, op,
extractBits(rewriter, adaptor.getInput()),
true);
1249 matchAndRewrite(
comb::ShlOp op, OpAdaptor adaptor,
1250 ConversionPatternRewriter &rewriter)
const override {
1251 auto width = op.getType().getIntOrFloatBitWidth();
1252 auto lhs = adaptor.getLhs();
1254 rewriter, op.getLoc(), adaptor.getRhs(), width,
1256 [&](int64_t index) {
1262 op.getLoc(), rewriter.getIntegerType(index), 0);
1265 [&](int64_t index) {
1266 assert(index < width &&
"index out of bounds");
1282 ConversionPatternRewriter &rewriter)
const override {
1283 auto width = op.getType().getIntOrFloatBitWidth();
1284 auto lhs = adaptor.getLhs();
1286 rewriter, op.getLoc(), adaptor.getRhs(), width,
1288 [&](int64_t index) {
1294 op.getLoc(), rewriter.getIntegerType(index), 0);
1297 [&](int64_t index) {
1298 assert(index < width &&
"index out of bounds");
1300 return rewriter.createOrFold<
comb::ExtractOp>(op.getLoc(), lhs, index,
1314 ConversionPatternRewriter &rewriter)
const override {
1315 auto width = op.getType().getIntOrFloatBitWidth();
1317 return rewriter.notifyMatchFailure(op.getLoc(),
1318 "i0 signed shift is unsupported");
1319 auto lhs = adaptor.getLhs();
1322 rewriter.createOrFold<
comb::ExtractOp>(op.getLoc(), lhs, width - 1, 1);
1327 rewriter, op.getLoc(), adaptor.getRhs(), width - 1,
1329 [&](int64_t index) {
1330 return rewriter.createOrFold<comb::ReplicateOp>(op.getLoc(), sign,
1334 [&](int64_t index) {
1335 return rewriter.createOrFold<
comb::ExtractOp>(op.getLoc(), lhs, index,
1351struct ConvertCombToSynthPass
1352 :
public impl::ConvertCombToSynthBase<ConvertCombToSynthPass> {
1353 void runOnOperation()
override;
1354 using ConvertCombToSynthBase<ConvertCombToSynthPass>::ConvertCombToSynthBase;
1360 uint32_t maxEmulationUnknownBits,
1364 CombAndOpConversion, CombMuxOpConversion, CombParityOpConversion,
1365 CombXorOpToSynthConversion,
1367 CombMulOpConversion, CombICmpOpConversion,
1369 CombShlOpConversion, CombShrUOpConversion, CombShrSOpConversion,
1371 CombLowerVariadicOp<AddOp>, CombLowerVariadicOp<MulOp>>(
1377 patterns.add(comb::convertSubToAdd);
1379 patterns.add<CombOrToAIGConversion, CombAddOpConversion>(
1381 synth::populateVariadicAndInverterLoweringPatterns(
patterns);
1384 synth::populateVariadicXorInverterLoweringPatterns(
patterns);
1387 patterns.add<CombDivUOpConversion, CombModUOpConversion, CombDivSOpConversion,
1388 CombModSOpConversion>(
patterns.getContext(),
1389 maxEmulationUnknownBits);
1392void ConvertCombToSynthPass::runOnOperation() {
1393 ConversionTarget target(getContext());
1396 target.addIllegalDialect<comb::CombDialect>();
1406 hw::AggregateConstantOp>();
1408 target.addLegalDialect<synth::SynthDialect>();
1410 target.addIllegalOp<synth::XorInverterOp>();
1413 if (!additionalLegalOps.empty())
1414 for (
const auto &opName : additionalLegalOps)
1415 target.addLegalOp(OperationName(opName, &getContext()));
1417 RewritePatternSet
patterns(&getContext());
1421 if (failed(mlir::applyPartialConversion(getOperation(), target,
1423 return signalPassFailure();
assert(baseType &&"element must be base type")
static SmallVector< Value > extractBits(OpBuilder &builder, Value val)
static Value createShiftLogic(ConversionPatternRewriter &rewriter, Location loc, Value shiftAmount, int64_t maxShiftAmount, llvm::function_ref< Value(int64_t)> getPadding, llvm::function_ref< Value(int64_t)> getExtract)
static APInt substitueMaskToValues(size_t width, llvm::SmallVectorImpl< ConstantOrValue > &constantOrValues, uint32_t mask)
static LogicalResult emulateBinaryOpForUnknownBits(ConversionPatternRewriter &rewriter, int64_t maxEmulationUnknownBits, Operation *op, llvm::function_ref< APInt(const APInt &, const APInt &)> emulate)
static int64_t getNumUnknownBitsAndPopulateValues(Value value, llvm::SmallVectorImpl< ConstantOrValue > &values)
static Value createMajorityFunction(OpBuilder &rewriter, Location loc, Value a, Value b, Value carry)
static Value extractOtherThanMSB(OpBuilder &builder, Value val)
static Value extractMSB(OpBuilder &builder, Value val)
static void populateCombToAIGConversionPatterns(RewritePatternSet &patterns, uint32_t maxEmulationUnknownBits, bool forceAIG)
static std::unique_ptr< Context > context
static std::optional< APSInt > getConstant(Attribute operand)
Determine the value of a constant operand for the sake of constant folding.
static Value lowerFullyAssociativeOp(Operation &op, OperandRange operands, SmallVector< Operation * > &newOps)
Lower a variadic fully-associative operation into an expression tree.
The InstanceGraph op interface, see InstanceGraphInterface.td for more details.
void replaceOpAndCopyNamehint(PatternRewriter &rewriter, Operation *op, Value newValue)
A wrapper of PatternRewriter::replaceOp to propagate "sv.namehint" attribute.