@@ -672,10 +672,10 @@ arrangeFreeFunctionLikeCall(CodeGenTypes &CGT, CodeGenModule &CGM,
672
672
addExtParameterInfosForCall (paramInfos, proto, numExtraRequiredArgs,
673
673
args.size ());
674
674
675
- // If we don't have a prototype at all, but we're supposed to
676
- // explicitly use the variadic convention for unprototyped calls,
677
- // treat all of the arguments as required but preserve the nominal
678
- // possibility of variadics.
675
+ // If we don't have a prototype at all, but we're supposed to
676
+ // explicitly use the variadic convention for unprototyped calls,
677
+ // treat all of the arguments as required but preserve the nominal
678
+ // possibility of variadics.
679
679
} else if (CGM.getTargetCodeGenInfo ().isNoProtoCallVariadic (
680
680
args, cast<FunctionNoProtoType>(fnType))) {
681
681
required = RequiredArgs (args.size ());
@@ -4061,7 +4061,7 @@ void CodeGenFunction::EmitFunctionEpilog(const CGFunctionInfo &FI,
4061
4061
if (results.size () == 1 ) {
4062
4062
RV = results[0 ];
4063
4063
4064
- // Otherwise, we need to make a first-class aggregate.
4064
+ // Otherwise, we need to make a first-class aggregate.
4065
4065
} else {
4066
4066
// Construct a return type that lacks padding elements.
4067
4067
llvm::Type *returnType = RetAI.getUnpaddedCoerceAndExpandType ();
@@ -4200,11 +4200,11 @@ void CodeGenFunction::EmitDelegateCallArg(CallArgList &args,
4200
4200
if (type->isReferenceType ()) {
4201
4201
args.add (RValue::get (Builder.CreateLoad (local)), type);
4202
4202
4203
- // In ARC, move out of consumed arguments so that the release cleanup
4204
- // entered by StartFunction doesn't cause an over-release. This isn't
4205
- // optimal -O0 code generation, but it should get cleaned up when
4206
- // optimization is enabled. This also assumes that delegate calls are
4207
- // performed exactly once for a set of arguments, but that should be safe.
4203
+ // In ARC, move out of consumed arguments so that the release cleanup
4204
+ // entered by StartFunction doesn't cause an over-release. This isn't
4205
+ // optimal -O0 code generation, but it should get cleaned up when
4206
+ // optimization is enabled. This also assumes that delegate calls are
4207
+ // performed exactly once for a set of arguments, but that should be safe.
4208
4208
} else if (getLangOpts ().ObjCAutoRefCount &&
4209
4209
param->hasAttr <NSConsumedAttr>() && type->isObjCRetainableType ()) {
4210
4210
llvm::Value *ptr = Builder.CreateLoad (local);
@@ -4213,8 +4213,8 @@ void CodeGenFunction::EmitDelegateCallArg(CallArgList &args,
4213
4213
Builder.CreateStore (null, local);
4214
4214
args.add (RValue::get (ptr), type);
4215
4215
4216
- // For the most part, we just need to load the alloca, except that
4217
- // aggregate r-values are actually pointers to temporaries.
4216
+ // For the most part, we just need to load the alloca, except that
4217
+ // aggregate r-values are actually pointers to temporaries.
4218
4218
} else {
4219
4219
args.add (convertTempToRValue (local, type, loc), type);
4220
4220
}
@@ -4306,7 +4306,7 @@ static void emitWriteback(CodeGenFunction &CGF,
4306
4306
// Release the old value.
4307
4307
CGF.EmitARCRelease (oldValue, srcLV.isARCPreciseLifetime ());
4308
4308
4309
- // Otherwise, we can just do a normal lvalue store.
4309
+ // Otherwise, we can just do a normal lvalue store.
4310
4310
} else {
4311
4311
CGF.EmitStoreThroughLValue (RValue::get (value), srcLV);
4312
4312
}
@@ -4347,7 +4347,7 @@ static void emitWritebackArg(CodeGenFunction &CGF, CallArgList &args,
4347
4347
if (const Expr *lvExpr = maybeGetUnaryAddrOfOperand (CRE->getSubExpr ())) {
4348
4348
srcLV = CGF.EmitLValue (lvExpr);
4349
4349
4350
- // Otherwise, just emit it as a scalar.
4350
+ // Otherwise, just emit it as a scalar.
4351
4351
} else {
4352
4352
Address srcAddr = CGF.EmitPointerWithAlignment (CRE->getSubExpr ());
4353
4353
0 commit comments