Index: llvm/trunk/docs/ProgrammersManual.rst =================================================================== --- llvm/trunk/docs/ProgrammersManual.rst +++ llvm/trunk/docs/ProgrammersManual.rst @@ -873,7 +873,7 @@ llvm/ADT/ilist_node.h ^^^^^^^^^^^^^^^^^^^^^ -``ilist_node`` implements a the forward and backward links that are expected +``ilist_node`` implements the forward and backward links that are expected by the ``ilist`` (and analogous containers) in the default manner. ``ilist_node``\ s are meant to be embedded in the node type ``T``, usually Index: llvm/trunk/include/llvm/Analysis/LazyCallGraph.h =================================================================== --- llvm/trunk/include/llvm/Analysis/LazyCallGraph.h +++ llvm/trunk/include/llvm/Analysis/LazyCallGraph.h @@ -537,7 +537,7 @@ static void *ID() { return (void *)&PassID; } - /// \brief Compute the \c LazyCallGraph for a the module \c M. + /// \brief Compute the \c LazyCallGraph for the module \c M. /// /// This just builds the set of entry points to the call graph. The rest is /// built lazily as it is walked. Index: llvm/trunk/include/llvm/Analysis/PostDominators.h =================================================================== --- llvm/trunk/include/llvm/Analysis/PostDominators.h +++ llvm/trunk/include/llvm/Analysis/PostDominators.h @@ -19,7 +19,7 @@ namespace llvm { /// PostDominatorTree Class - Concrete subclass of DominatorTree that is used to -/// compute the a post-dominator tree. +/// compute the post-dominator tree. /// struct PostDominatorTree : public FunctionPass { static char ID; // Pass identification, replacement for typeid Index: llvm/trunk/include/llvm/Analysis/TargetTransformInfo.h =================================================================== --- llvm/trunk/include/llvm/Analysis/TargetTransformInfo.h +++ llvm/trunk/include/llvm/Analysis/TargetTransformInfo.h @@ -183,7 +183,7 @@ /// should probably move to simpler cost metrics using the above. /// Alternatively, we could split the cost interface into distinct code-size /// and execution-speed costs. This would allow modelling the core of this - /// query more accurately as the a call is a single small instruction, but + /// query more accurately as a call is a single small instruction, but /// incurs significant execution cost. virtual bool isLoweredToCall(const Function *F) const; Index: llvm/trunk/include/llvm/CodeGen/MachinePostDominators.h =================================================================== --- llvm/trunk/include/llvm/CodeGen/MachinePostDominators.h +++ llvm/trunk/include/llvm/CodeGen/MachinePostDominators.h @@ -22,7 +22,7 @@ /// /// PostDominatorTree Class - Concrete subclass of DominatorTree that is used -/// to compute the a post-dominator tree. +/// to compute the post-dominator tree. /// struct MachinePostDominatorTree : public MachineFunctionPass { private: Index: llvm/trunk/include/llvm/CodeGen/SelectionDAG.h =================================================================== --- llvm/trunk/include/llvm/CodeGen/SelectionDAG.h +++ llvm/trunk/include/llvm/CodeGen/SelectionDAG.h @@ -749,7 +749,7 @@ SDValue SV, unsigned Align); /// getAtomicCmpSwap - Gets a node for an atomic cmpxchg op. There are two - /// valid Opcodes. ISD::ATOMIC_CMO_SWAP produces a the value loaded and a + /// valid Opcodes. ISD::ATOMIC_CMO_SWAP produces the value loaded and a /// chain result. ISD::ATOMIC_CMP_SWAP_WITH_SUCCESS produces the value loaded, /// a success flag (initially i1), and a chain. SDValue getAtomicCmpSwap(unsigned Opcode, SDLoc dl, EVT MemVT, SDVTList VTs, Index: llvm/trunk/include/llvm/IR/IRBuilder.h =================================================================== --- llvm/trunk/include/llvm/IR/IRBuilder.h +++ llvm/trunk/include/llvm/IR/IRBuilder.h @@ -446,7 +446,7 @@ /// The first template argument handles whether or not to preserve names in the /// final instruction output. This defaults to on. The second template argument /// specifies a class to use for creating constants. This defaults to creating -/// minimally folded constants. The fourth template argument allows clients to +/// minimally folded constants. The third template argument allows clients to /// specify custom insertion hooks that are called on every newly created /// insertion. template {-1,+,3,+,2} if (AR->isAffine() && Index: llvm/trunk/lib/AsmParser/LLParser.cpp =================================================================== --- llvm/trunk/lib/AsmParser/LLParser.cpp +++ llvm/trunk/lib/AsmParser/LLParser.cpp @@ -3337,7 +3337,7 @@ if (isDefine) return false; - // Check the a declaration has no block address forward references. + // Check the declaration has no block address forward references. ValID ID; if (FunctionName.empty()) { ID.Kind = ValID::t_GlobalID; Index: llvm/trunk/lib/CodeGen/AsmPrinter/DIEHash.cpp =================================================================== --- llvm/trunk/lib/CodeGen/AsmPrinter/DIEHash.cpp +++ llvm/trunk/lib/CodeGen/AsmPrinter/DIEHash.cpp @@ -261,7 +261,7 @@ return; } - // otherwise, b) use the letter 'T' as a the marker, ... + // otherwise, b) use the letter 'T' as the marker, ... addULEB128('T'); addULEB128(Attribute); Index: llvm/trunk/lib/CodeGen/StackColoring.cpp =================================================================== --- llvm/trunk/lib/CodeGen/StackColoring.cpp +++ llvm/trunk/lib/CodeGen/StackColoring.cpp @@ -228,7 +228,7 @@ unsigned StackColoring::collectMarkers(unsigned NumSlot) { unsigned MarkersFound = 0; // Scan the function to find all lifetime markers. - // NOTE: We use the a reverse-post-order iteration to ensure that we obtain a + // NOTE: We use a reverse-post-order iteration to ensure that we obtain a // deterministic numbering, and because we'll need a post-order iteration // later for solving the liveness dataflow problem. for (MachineBasicBlock *MBB : depth_first(MF)) { Index: llvm/trunk/lib/ExecutionEngine/JIT/JITMemoryManager.cpp =================================================================== --- llvm/trunk/lib/ExecutionEngine/JIT/JITMemoryManager.cpp +++ llvm/trunk/lib/ExecutionEngine/JIT/JITMemoryManager.cpp @@ -680,7 +680,7 @@ bool DefaultJITMemoryManager::CheckInvariants(std::string &ErrorStr) { raw_string_ostream Err(ErrorStr); - // Construct a the set of FreeRangeHeader pointers so we can query it + // Construct the set of FreeRangeHeader pointers so we can query it // efficiently. llvm::SmallPtrSet FreeHdrSet; FreeRangeHeader* FreeHead = FreeMemoryList; Index: llvm/trunk/lib/IR/DiagnosticPrinter.cpp =================================================================== --- llvm/trunk/lib/IR/DiagnosticPrinter.cpp +++ llvm/trunk/lib/IR/DiagnosticPrinter.cpp @@ -7,7 +7,7 @@ // //===----------------------------------------------------------------------===// // -// This file defines the a diagnostic printer relying on raw_ostream. +// This file defines a diagnostic printer relying on raw_ostream. // //===----------------------------------------------------------------------===// Index: llvm/trunk/lib/Target/AArch64/AArch64ISelDAGToDAG.cpp =================================================================== --- llvm/trunk/lib/Target/AArch64/AArch64ISelDAGToDAG.cpp +++ llvm/trunk/lib/Target/AArch64/AArch64ISelDAGToDAG.cpp @@ -303,7 +303,7 @@ /// \brief Determine wether it is worth to fold V into an extended register. bool AArch64DAGToDAGISel::isWorthFolding(SDValue V) const { - // it hurts if the a value is used at least twice, unless we are optimizing + // it hurts if the value is used at least twice, unless we are optimizing // for code size. if (ForCodeSize || V.hasOneUse()) return true; Index: llvm/trunk/lib/Target/AArch64/AArch64ISelLowering.h =================================================================== --- llvm/trunk/lib/Target/AArch64/AArch64ISelLowering.h +++ llvm/trunk/lib/Target/AArch64/AArch64ISelLowering.h @@ -199,8 +199,7 @@ public: explicit AArch64TargetLowering(TargetMachine &TM); - /// Selects the correct CCAssignFn for a the given CallingConvention - /// value. + /// Selects the correct CCAssignFn for a given CallingConvention value. CCAssignFn *CCAssignFnForCall(CallingConv::ID CC, bool IsVarArg) const; /// computeKnownBitsForTargetNode - Determine which of the bits specified in Index: llvm/trunk/lib/Target/AArch64/AArch64ISelLowering.cpp =================================================================== --- llvm/trunk/lib/Target/AArch64/AArch64ISelLowering.cpp +++ llvm/trunk/lib/Target/AArch64/AArch64ISelLowering.cpp @@ -1781,8 +1781,7 @@ #include "AArch64GenCallingConv.inc" -/// Selects the correct CCAssignFn for a the given CallingConvention -/// value. +/// Selects the correct CCAssignFn for a given CallingConvention value. CCAssignFn *AArch64TargetLowering::CCAssignFnForCall(CallingConv::ID CC, bool IsVarArg) const { switch (CC) { Index: llvm/trunk/lib/Target/ARM/ARMBaseInstrInfo.cpp =================================================================== --- llvm/trunk/lib/Target/ARM/ARMBaseInstrInfo.cpp +++ llvm/trunk/lib/Target/ARM/ARMBaseInstrInfo.cpp @@ -2889,7 +2889,7 @@ // FIXME: The current MachineInstr design does not support relying on machine // mem operands to determine the width of a memory access. Instead, we expect // the target to provide this information based on the instruction opcode and -// operands. However, using MachineMemOperand is a the best solution now for +// operands. However, using MachineMemOperand is the best solution now for // two reasons: // // 1) getNumMicroOps tries to infer LDM memory width from the total number of MI Index: llvm/trunk/lib/Target/Mips/Mips16InstrInfo.td =================================================================== --- llvm/trunk/lib/Target/Mips/Mips16InstrInfo.td +++ llvm/trunk/lib/Target/Mips/Mips16InstrInfo.td @@ -1771,9 +1771,9 @@ // // For constants, llvm transforms this to: -// x > (k -1) and then reverses the operands to use setlt. So this pattern +// x > (k - 1) and then reverses the operands to use setlt. So this pattern // is not used now by the compiler. (Presumably checking that k-1 does not -// overflow). The compiler never uses this at a the current time, due to +// overflow). The compiler never uses this at the current time, due to // other optimizations. // //def: Mips16Pat Index: llvm/trunk/test/CodeGen/X86/2013-10-14-FastISel-incorrect-vreg.ll =================================================================== --- llvm/trunk/test/CodeGen/X86/2013-10-14-FastISel-incorrect-vreg.ll +++ llvm/trunk/test/CodeGen/X86/2013-10-14-FastISel-incorrect-vreg.ll @@ -6,7 +6,7 @@ ; we may reference variables that were not live across basic blocks ; resulting in undefined virtual registers. ; -; In this example, this is illustrated by a the spill/reload of the +; In this example, this is illustrated by a spill/reload of the ; LOADED_PTR_SLOT. ; ; Before this patch, the compiler was accessing two different spill Index: llvm/trunk/test/MC/ELF/comdat.s =================================================================== --- llvm/trunk/test/MC/ELF/comdat.s +++ llvm/trunk/test/MC/ELF/comdat.s @@ -1,6 +1,6 @@ // RUN: llvm-mc -filetype=obj -triple x86_64-pc-linux-gnu %s -o - | llvm-readobj -s -t | FileCheck %s -// Test that we produce the group sections and that they are a the beginning +// Test that we produce the group sections and that they are at the beginning // of the file. // CHECK: Section { Index: llvm/trunk/utils/TableGen/DAGISelMatcherOpt.cpp =================================================================== --- llvm/trunk/utils/TableGen/DAGISelMatcherOpt.cpp +++ llvm/trunk/utils/TableGen/DAGISelMatcherOpt.cpp @@ -185,7 +185,7 @@ /// Conceptually, we'd like to sink these predicates all the way to the last /// matcher predicate in the series. However, it turns out that some /// ComplexPatterns have side effects on the graph, so we really don't want to -/// run a the complex pattern if the pattern predicate will fail. For this +/// run a complex pattern if the pattern predicate will fail. For this /// reason, we refuse to sink the pattern predicate past a ComplexPattern. /// static void SinkPatternPredicates(std::unique_ptr &MatcherPtr) {