@@ -672,10 +672,10 @@ arrangeFreeFunctionLikeCall(CodeGenTypes &CGT, CodeGenModule &CGM,
672672 addExtParameterInfosForCall (paramInfos, proto, numExtraRequiredArgs,
673673 args.size ());
674674
675- // If we don't have a prototype at all, but we're supposed to
676- // explicitly use the variadic convention for unprototyped calls,
677- // treat all of the arguments as required but preserve the nominal
678- // possibility of variadics.
675+ // If we don't have a prototype at all, but we're supposed to
676+ // explicitly use the variadic convention for unprototyped calls,
677+ // treat all of the arguments as required but preserve the nominal
678+ // possibility of variadics.
679679 } else if (CGM.getTargetCodeGenInfo ().isNoProtoCallVariadic (
680680 args, cast<FunctionNoProtoType>(fnType))) {
681681 required = RequiredArgs (args.size ());
@@ -4061,7 +4061,7 @@ void CodeGenFunction::EmitFunctionEpilog(const CGFunctionInfo &FI,
40614061 if (results.size () == 1 ) {
40624062 RV = results[0 ];
40634063
4064- // Otherwise, we need to make a first-class aggregate.
4064+ // Otherwise, we need to make a first-class aggregate.
40654065 } else {
40664066 // Construct a return type that lacks padding elements.
40674067 llvm::Type *returnType = RetAI.getUnpaddedCoerceAndExpandType ();
@@ -4200,11 +4200,11 @@ void CodeGenFunction::EmitDelegateCallArg(CallArgList &args,
42004200 if (type->isReferenceType ()) {
42014201 args.add (RValue::get (Builder.CreateLoad (local)), type);
42024202
4203- // In ARC, move out of consumed arguments so that the release cleanup
4204- // entered by StartFunction doesn't cause an over-release. This isn't
4205- // optimal -O0 code generation, but it should get cleaned up when
4206- // optimization is enabled. This also assumes that delegate calls are
4207- // performed exactly once for a set of arguments, but that should be safe.
4203+ // In ARC, move out of consumed arguments so that the release cleanup
4204+ // entered by StartFunction doesn't cause an over-release. This isn't
4205+ // optimal -O0 code generation, but it should get cleaned up when
4206+ // optimization is enabled. This also assumes that delegate calls are
4207+ // performed exactly once for a set of arguments, but that should be safe.
42084208 } else if (getLangOpts ().ObjCAutoRefCount &&
42094209 param->hasAttr <NSConsumedAttr>() && type->isObjCRetainableType ()) {
42104210 llvm::Value *ptr = Builder.CreateLoad (local);
@@ -4213,8 +4213,8 @@ void CodeGenFunction::EmitDelegateCallArg(CallArgList &args,
42134213 Builder.CreateStore (null, local);
42144214 args.add (RValue::get (ptr), type);
42154215
4216- // For the most part, we just need to load the alloca, except that
4217- // aggregate r-values are actually pointers to temporaries.
4216+ // For the most part, we just need to load the alloca, except that
4217+ // aggregate r-values are actually pointers to temporaries.
42184218 } else {
42194219 args.add (convertTempToRValue (local, type, loc), type);
42204220 }
@@ -4306,7 +4306,7 @@ static void emitWriteback(CodeGenFunction &CGF,
43064306 // Release the old value.
43074307 CGF.EmitARCRelease (oldValue, srcLV.isARCPreciseLifetime ());
43084308
4309- // Otherwise, we can just do a normal lvalue store.
4309+ // Otherwise, we can just do a normal lvalue store.
43104310 } else {
43114311 CGF.EmitStoreThroughLValue (RValue::get (value), srcLV);
43124312 }
@@ -4347,7 +4347,7 @@ static void emitWritebackArg(CodeGenFunction &CGF, CallArgList &args,
43474347 if (const Expr *lvExpr = maybeGetUnaryAddrOfOperand (CRE->getSubExpr ())) {
43484348 srcLV = CGF.EmitLValue (lvExpr);
43494349
4350- // Otherwise, just emit it as a scalar.
4350+ // Otherwise, just emit it as a scalar.
43514351 } else {
43524352 Address srcAddr = CGF.EmitPointerWithAlignment (CRE->getSubExpr ());
43534353
0 commit comments