@@ -117,8 +117,8 @@ class CoroCloner {
117
117
// / Create a cloner for a switch lowering.
118
118
CoroCloner (Function &OrigF, const Twine &Suffix, coro::Shape &Shape,
119
119
Kind FKind)
120
- : OrigF(OrigF), NewF(nullptr ), Suffix(Suffix), Shape(Shape),
121
- FKind (FKind), Builder(OrigF.getContext()) {
120
+ : OrigF(OrigF), NewF(nullptr ), Suffix(Suffix), Shape(Shape), FKind(FKind ),
121
+ Builder (OrigF.getContext()) {
122
122
assert (Shape.ABI == coro::ABI::Switch);
123
123
}
124
124
@@ -170,8 +170,7 @@ class CoroCloner {
170
170
static void maybeFreeRetconStorage (IRBuilder<> &Builder,
171
171
const coro::Shape &Shape, Value *FramePtr,
172
172
CallGraph *CG) {
173
- assert (Shape.ABI == coro::ABI::Retcon ||
174
- Shape.ABI == coro::ABI::RetconOnce);
173
+ assert (Shape.ABI == coro::ABI::Retcon || Shape.ABI == coro::ABI::RetconOnce);
175
174
if (Shape.RetconLowering .IsFrameInlineInStorage )
176
175
return ;
177
176
@@ -208,12 +207,17 @@ static bool replaceCoroEndAsync(AnyCoroEndInst *End) {
208
207
// Insert the return instruction.
209
208
Builder.SetInsertPoint (End);
210
209
Builder.CreateRetVoid ();
210
+ InlineFunctionInfo FnInfo;
211
211
212
212
// Remove the rest of the block, by splitting it into an unreachable block.
213
213
auto *BB = End->getParent ();
214
214
BB->splitBasicBlock (End);
215
215
BB->getTerminator ()->eraseFromParent ();
216
216
217
+ auto InlineRes = InlineFunction (*MustTailCall, FnInfo);
218
+ assert (InlineRes.isSuccess () && " Expected inlining to succeed" );
219
+ (void )InlineRes;
220
+
217
221
// We have cleaned up the coro.end block above.
218
222
return false ;
219
223
}
@@ -264,7 +268,7 @@ static void replaceFallthroughCoroEnd(AnyCoroEndInst *End,
264
268
265
269
if (auto *RetStructTy = dyn_cast<StructType>(RetTy)) {
266
270
assert (RetStructTy->getNumElements () == NumReturns &&
267
- " numbers of returns should match resume function singature" );
271
+ " numbers of returns should match resume function singature" );
268
272
Value *ReturnValue = UndefValue::get (RetStructTy);
269
273
unsigned Idx = 0 ;
270
274
for (Value *RetValEl : CoroResults->return_values ())
@@ -277,7 +281,8 @@ static void replaceFallthroughCoroEnd(AnyCoroEndInst *End,
277
281
assert (NumReturns == 1 );
278
282
Builder.CreateRet (*CoroResults->retval_begin ());
279
283
}
280
- CoroResults->replaceAllUsesWith (ConstantTokenNone::get (CoroResults->getContext ()));
284
+ CoroResults->replaceAllUsesWith (
285
+ ConstantTokenNone::get (CoroResults->getContext ()));
281
286
CoroResults->eraseFromParent ();
282
287
break ;
283
288
}
@@ -291,7 +296,7 @@ static void replaceFallthroughCoroEnd(AnyCoroEndInst *End,
291
296
auto RetTy = Shape.getResumeFunctionType ()->getReturnType ();
292
297
auto RetStructTy = dyn_cast<StructType>(RetTy);
293
298
PointerType *ContinuationTy =
294
- cast<PointerType>(RetStructTy ? RetStructTy->getElementType (0 ) : RetTy);
299
+ cast<PointerType>(RetStructTy ? RetStructTy->getElementType (0 ) : RetTy);
295
300
296
301
Value *ReturnValue = ConstantPointerNull::get (ContinuationTy);
297
302
if (RetStructTy) {
@@ -480,11 +485,12 @@ void CoroCloner::replaceRetconOrAsyncSuspendUses() {
480
485
Shape.ABI == coro::ABI::Async);
481
486
482
487
auto NewS = VMap[ActiveSuspend];
483
- if (NewS->use_empty ()) return ;
488
+ if (NewS->use_empty ())
489
+ return ;
484
490
485
491
// Copy out all the continuation arguments after the buffer pointer into
486
492
// an easily-indexed data structure for convenience.
487
- SmallVector<Value*, 8 > Args;
493
+ SmallVector<Value *, 8 > Args;
488
494
// The async ABI includes all arguments -- including the first argument.
489
495
bool IsAsyncABI = Shape.ABI == coro::ABI::Async;
490
496
for (auto I = IsAsyncABI ? NewF->arg_begin () : std::next (NewF->arg_begin ()),
@@ -511,7 +517,8 @@ void CoroCloner::replaceRetconOrAsyncSuspendUses() {
511
517
}
512
518
513
519
// If we have no remaining uses, we're done.
514
- if (NewS->use_empty ()) return ;
520
+ if (NewS->use_empty ())
521
+ return ;
515
522
516
523
// Otherwise, we need to create an aggregate.
517
524
Value *Agg = PoisonValue::get (NewS->getType ());
@@ -549,7 +556,8 @@ void CoroCloner::replaceCoroSuspends() {
549
556
550
557
for (AnyCoroSuspendInst *CS : Shape.CoroSuspends ) {
551
558
// The active suspend was handled earlier.
552
- if (CS == ActiveSuspend) continue ;
559
+ if (CS == ActiveSuspend)
560
+ continue ;
553
561
554
562
auto *MappedCS = cast<AnyCoroSuspendInst>(VMap[CS]);
555
563
MappedCS->replaceAllUsesWith (SuspendResult);
@@ -707,7 +715,7 @@ void CoroCloner::replaceEntryBlock() {
707
715
// In switch-lowering, we built a resume-entry block in the original
708
716
// function. Make the entry block branch to this.
709
717
auto *SwitchBB =
710
- cast<BasicBlock>(VMap[Shape.SwitchLowering .ResumeEntryBlock ]);
718
+ cast<BasicBlock>(VMap[Shape.SwitchLowering .ResumeEntryBlock ]);
711
719
Builder.CreateBr (SwitchBB);
712
720
break ;
713
721
}
@@ -1055,7 +1063,7 @@ void CoroCloner::create() {
1055
1063
// to suppress deallocation code.
1056
1064
if (Shape.ABI == coro::ABI::Switch)
1057
1065
coro::replaceCoroFree (cast<CoroIdInst>(VMap[Shape.CoroBegin ->getId ()]),
1058
- /* Elide=*/ FKind == CoroCloner::Kind::SwitchCleanup);
1066
+ /* Elide=*/ FKind == CoroCloner::Kind::SwitchCleanup);
1059
1067
}
1060
1068
1061
1069
static void updateAsyncFuncPointerContextSize (coro::Shape &Shape) {
@@ -1842,8 +1850,13 @@ static void splitAsyncCoroutine(Function &F, coro::Shape &Shape,
1842
1850
SmallVector<Value *, 8 > Args (Suspend->args ());
1843
1851
auto FnArgs = ArrayRef<Value *>(Args).drop_front (
1844
1852
CoroSuspendAsyncInst::MustTailCallFuncArg + 1 );
1845
- coro::createMustTailCall (Suspend->getDebugLoc (), Fn, TTI, FnArgs, Builder);
1853
+ auto *TailCall = coro::createMustTailCall (Suspend->getDebugLoc (), Fn, TTI,
1854
+ FnArgs, Builder);
1846
1855
Builder.CreateRetVoid ();
1856
+ InlineFunctionInfo FnInfo;
1857
+ auto InlineRes = InlineFunction (*TailCall, FnInfo);
1858
+ assert (InlineRes.isSuccess () && " Expected inlining to succeed" );
1859
+ (void )InlineRes;
1847
1860
1848
1861
// Replace the lvm.coro.async.resume intrisic call.
1849
1862
replaceAsyncResumeFunction (Suspend, Continuation);
@@ -1860,8 +1873,7 @@ static void splitAsyncCoroutine(Function &F, coro::Shape &Shape,
1860
1873
1861
1874
static void splitRetconCoroutine (Function &F, coro::Shape &Shape,
1862
1875
SmallVectorImpl<Function *> &Clones) {
1863
- assert (Shape.ABI == coro::ABI::Retcon ||
1864
- Shape.ABI == coro::ABI::RetconOnce);
1876
+ assert (Shape.ABI == coro::ABI::Retcon || Shape.ABI == coro::ABI::RetconOnce);
1865
1877
assert (Clones.empty ());
1866
1878
1867
1879
// Reset various things that the optimizer might have decided it
@@ -1887,7 +1899,7 @@ static void splitRetconCoroutine(Function &F, coro::Shape &Shape,
1887
1899
// FIXME: pass the required alignment
1888
1900
RawFramePtr = Shape.emitAlloc (Builder, Builder.getInt64 (Size ), nullptr );
1889
1901
RawFramePtr =
1890
- Builder.CreateBitCast (RawFramePtr, Shape.CoroBegin ->getType ());
1902
+ Builder.CreateBitCast (RawFramePtr, Shape.CoroBegin ->getType ());
1891
1903
1892
1904
// Stash the allocated frame pointer in the continuation storage.
1893
1905
Builder.CreateStore (RawFramePtr, Id->getStorage ());
@@ -1927,8 +1939,8 @@ static void splitRetconCoroutine(Function &F, coro::Shape &Shape,
1927
1939
// Create the unified return block.
1928
1940
if (!ReturnBB) {
1929
1941
// Place it before the first suspend.
1930
- ReturnBB = BasicBlock::Create (F. getContext (), " coro.return " , &F,
1931
- NewSuspendBB);
1942
+ ReturnBB =
1943
+ BasicBlock::Create (F. getContext (), " coro.return " , &F, NewSuspendBB);
1932
1944
Shape.RetconLowering .ReturnBlock = ReturnBB;
1933
1945
1934
1946
IRBuilder<> Builder (ReturnBB);
@@ -1942,8 +1954,8 @@ static void splitRetconCoroutine(Function &F, coro::Shape &Shape,
1942
1954
1943
1955
// Next, all the directly-yielded values.
1944
1956
for (auto *ResultTy : Shape.getRetconResultTypes ())
1945
- ReturnPHIs.push_back (Builder. CreatePHI (ResultTy,
1946
- Shape.CoroSuspends .size ()));
1957
+ ReturnPHIs.push_back (
1958
+ Builder. CreatePHI (ResultTy, Shape.CoroSuspends .size ()));
1947
1959
1948
1960
// Build the return value.
1949
1961
auto RetTy = F.getReturnType ();
@@ -1952,9 +1964,9 @@ static void splitRetconCoroutine(Function &F, coro::Shape &Shape,
1952
1964
// We can't rely on the types matching up because that type would
1953
1965
// have to be infinite.
1954
1966
auto CastedContinuationTy =
1955
- (ReturnPHIs.size () == 1 ? RetTy : RetTy->getStructElementType (0 ));
1967
+ (ReturnPHIs.size () == 1 ? RetTy : RetTy->getStructElementType (0 ));
1956
1968
auto *CastedContinuation =
1957
- Builder.CreateBitCast (ReturnPHIs[0 ], CastedContinuationTy);
1969
+ Builder.CreateBitCast (ReturnPHIs[0 ], CastedContinuationTy);
1958
1970
1959
1971
Value *RetV;
1960
1972
if (ReturnPHIs.size () == 1 ) {
@@ -1988,17 +2000,18 @@ static void splitRetconCoroutine(Function &F, coro::Shape &Shape,
1988
2000
}
1989
2001
1990
2002
namespace {
1991
- class PrettyStackTraceFunction : public PrettyStackTraceEntry {
1992
- Function &F;
1993
- public:
1994
- PrettyStackTraceFunction (Function &F) : F(F) {}
1995
- void print (raw_ostream &OS) const override {
1996
- OS << " While splitting coroutine " ;
1997
- F.printAsOperand (OS, /* print type*/ false , F.getParent ());
1998
- OS << " \n " ;
1999
- }
2000
- };
2001
- }
2003
+ class PrettyStackTraceFunction : public PrettyStackTraceEntry {
2004
+ Function &F;
2005
+
2006
+ public:
2007
+ PrettyStackTraceFunction (Function &F) : F(F) {}
2008
+ void print (raw_ostream &OS) const override {
2009
+ OS << " While splitting coroutine " ;
2010
+ F.printAsOperand (OS, /* print type*/ false , F.getParent ());
2011
+ OS << " \n " ;
2012
+ }
2013
+ };
2014
+ } // namespace
2002
2015
2003
2016
static coro::Shape
2004
2017
splitCoroutine (Function &F, SmallVectorImpl<Function *> &Clones,
0 commit comments