Details
Diff Detail
- Repository
- rG LLVM Github Monorepo
Event Timeline
mlir/include/mlir/Dialect/SparseTensor/Utils/Merger.h | ||
---|---|---|
446 | What do you mean here? I'm guessing this should be "level", but I'm not sure what you mean by "favor(ing) constant levels" | |
mlir/lib/Dialect/SparseTensor/Transforms/LoopEmitter.h | ||
138 | This should remain "/// Exits" | |
239–260 | Can you make this class "final" (ditto for the LoopInfo). You don't need/use subclassing here, and marking it final helps the compiler generate more efficient code | |
243 | I'm guessing this should be LoopOrd. If not, then what is it? | |
315 | This should be Level lvl (I'm pretty sure) | |
417 | Should this be LoopOrd? If not, then what is it? | |
439 | What is this supposed to be: LoopOrd, LoopId, other? |
mlir/lib/Dialect/SparseTensor/Transforms/LoopEmitter.cpp | ||
---|---|---|
37–38 | I'd prefer these be defined as static inline functions rather than as macros, since that gives better type-safety and compiler error messages. If you need to use CMPI at several different types, then just use a template. | |
45–47 | You should just use ValueRange::getTypes() | |
312–315 | Please keep this as Level l | |
425 | Please use TensorId here. I am just about to upload the CLs that make that into a newtype, so it's important to use the correct type instead of just using size_t/unsigned everywhere | |
426 | Please define a dyn_cast variant of the getSparseTensorType function, and use that here and everywhere else. The SparseTensorType was created specifically to help avoid several code legibility and correctness concerns, so you should be using it everywhere possible. | |
430 | You should be using SparseTensorType::getLevelRank here, since it is specifically the level-rank you want not the dim-rank | |
431 | Please use Level for all levels. Even though it's just a typedef for now, I will be converting it to a proper type in the near future, so you should use the correct type rather than just using unsigned everywhere | |
472 | I think it'd be clearer to just use const auto & here | |
634 | Use "l" or "lvl" here. The name "d" is reserved for things of Dimension type, whereas this has Level type. | |
810 | Why remove the const? It's clearer to know when local variables will never change | |
817 | Please don't undo my factoring this out into a local variable. The condition is much easier to read when (1) it's all on one line, and (2) avoids repeating common expressions which forces the reader to double check if they are indeed the same or not. | |
1194 | It would be clearer to use break in the then-branch. That keeps you from needing to indent the else-branch (which is very long), and helps the reader avoid needing to check to see if there's something else after the else-branch. |
mlir/lib/Dialect/SparseTensor/Transforms/LoopEmitter.cpp | ||
---|---|---|
37–38 | I will stick with this. The reason I use macro is that I want to avoid typing builder and loc. | |
817 | Okay, it is a mistake I made during rebasing. | |
mlir/lib/Dialect/SparseTensor/Transforms/LoopEmitter.h | ||
417 | This is unsigned index to another array (not the loop sequence). I will stick with this. | |
439 | This is an unsigned counter. |
address comments + fix rebase mistakes.
mlir/include/mlir/Dialect/SparseTensor/Utils/Merger.h | ||
---|---|---|
446 | I mean to pick the a static known dimension size instead of dynamic ones if there are multiple candidate. (i.e., DimOp folds to constant value). It might lead to a slightly better code. | |
mlir/lib/Dialect/SparseTensor/Transforms/LoopEmitter.cpp | ||
810 | rebase mistake. | |
1194 | I found else is easier to follow, because the control flow is more straight forward. |
mlir/lib/Dialect/SparseTensor/Transforms/LoopEmitter.cpp | ||
---|---|---|
43 | Please don't undo this variable naming. The "mem" matches other places, and avoids confusion about whether "ptr" means MemRef vs the old "pointer" (now "position") vs llvm-pointers vs... | |
251–256 | You should use the numTensors variable instead of calling tensors.size() repeatedly. (This is for code clarity rather than performance reasons) | |
318–322 | It would be clearer to combine these together. Also, it would be clearer to use continue rather than extra indentation for the conditional. Putting those together, maybe use something like if (depends == 0) continue; assert(!reassoc); sliceSizes[... | |
mlir/lib/Dialect/SparseTensor/Transforms/LoopEmitter.h | ||
138 | You want to keep the triple-slash "///", since that's what the tooling uses for generating the API documentation |
mlir/include/mlir/Dialect/SparseTensor/Utils/Merger.h | ||
---|---|---|
478 | That's the wrong bound to assert; you should instead compare against levelToDependentIdx[t].size() or maxLvlRank. For correctness you should also assert(i < numLoops && t < numTensors). If you rebase over D146684 to get the assertion helpers, then the full assertion would be assert(isValidLevel(t, lvl) && isValidLoopId(i) && !loopToDependencies[i][t].has_value()). The nice thing about those assertion helpers is that it gives the correct bound for the level (rather than falling back to maxLvlRank), and also guards against the case where i == kInvalidId. | |
479 | Isn't it redundant to store the level-type in loopToDependencies, since it's already stored in lvlTypes? That is, afaict the following code snippet should always return successfully (assuming i and t are valid): if (const auto dep = loopToDependencies[i][t]) { const Level depLvl = (*dep).first; const auto depOptLoop = lvlToLoop[t][depLvl]; assert(depOptLoop); const LoopId depLoop = *depOptLoop; assert(lvlTypes[t][depLoop] == (*dep).second); } Assuming that's correct, then you shouldn't store the level-type in loopToDependencies because it's redundant information. Or if you absolutely must store the redundant copy for some reason, then you need to verify that the level-type agrees with the one in lvlTypes (or if the one in lvlTypes is undefined, then you need to store the dlt parameter there too; and conversely you need to adjust the setLevelAndType method to also verify consistency with loopToDependencies). I agree that it's rather convoluted to need to say lvlTypes[t][*(lvlToLoop[t][*(loopToDependencies[i][t])])], but the only solution to that is to reconsider the design of all the fields of the Merger class. For example, if we had lvlTypes : (TensorId, Level) -> LevelType instead of the current lvlTypes : (TensorId, LoopId) -> LevelType, then you wouldn't need to use lvlToLoop there. Of course, you'd have to redo the rest of the Merger code to make that work (which may end up inserting more loopToLevel uses than however many levelToLoop uses it removes). Or a different design would be to combine lvlToLoop and lvlTypes into a single lvlInfo : (TensorId, Level) -> (LevelType, optional<LoopId>); of course that would require making different changes to the rest of the Merger code. In any case, I think it would be wise to wait for D146693 to land before trying to make any of these changes, since the newtypes of that CL will greatly simplify the process of rearranging all these vectors. | |
651 | This should have been named levelToDependentLoop, since we don't use "idx" in this file anymore because it causes too much confusion. I mentioned this in the previous CL that introduced this field, but you landed the CL without fixing it | |
mlir/lib/Dialect/SparseTensor/Transforms/LoopEmitter.h | ||
228 | This comment should explain how exactly the slicedTids differs from the other tids. Also, is the same tensor allowed to occur in both fields? If so, then what does that mean? and, can the same level of the same tensor occur on both of the corresponding fields? | |
229–233 | I think it'd be better to combine these all into a single const SmallVector<LoopLevelInfo> where struct LoopLevelInfo final { TensorId tid; Level lvl; bool isSlice; bool isReduced; }; —assuming it's okay to combine the original (tids,lvls) with the new (slicedTids,slicedLvls,sliceReduced) into a single vector/set. If that's not okay for some reason, then I still think it'd be good to use a single const SmallVector<LoopSlicedLevelInfo> field for the new stuff. Using AOS will ensure that there's the right number of all the things that should correspond, as well as keeping the corresponding things close together. Plus it'll make it easier to add additional fields in the future as needed. | |
230 | "levels" | |
232 | This comment is wrong for this field | |
240 | "...do not need to actually create a sparse..." | |
241 | "...only need to maintain the..." | |
242 | Is this actually the full MemRef of coordinates for all levels? If so then it should be "minCoords" (with an "s"). Whereas if it's just a single coordinate for the given level, then it should be "minCrd". | |
250 | Given the comment, I think this would be better named something like "isFirstSlice" or "isInitialSlice". | |
252 | If the value is just a single coordinate, then this should also be singular. |
mlir/lib/Dialect/SparseTensor/Transforms/LoopEmitter.h | ||
---|---|---|
246 | Given our discussion of boolean blindness, this assertion suggests that the two parameters should be combined into a single std::optional<std::pair<Level, Value>> parameter. (Albeit you'll still want to assert you didn't get a null Value.) If that combined parameter doesn't work, why not? The only other thing that would be consistent with the assertion is union{ non-null-Value; struct{non-null-Value; Level}} but that's equivalent to struct{non-null-Value; optional<Level>}, so if that's what you want then you should assert(minCoord) instead. | |
386–388 | This description doesn't make sense to me. Do you have a design doc that explains what exactly you mean by "slice" in this context, and explains how/why "reducing d0+d1+d2" translates into needing the two slices you mention? | |
392 | Either "number of constraints needed to..." or "number of constraints that are needed to..." | |
393 | "level" | |
398 | That should be "A[i+j] => A[i+2]" to make it clear that it dereives from "j => 2". |
mlir/include/mlir/Dialect/SparseTensor/Utils/Merger.h | ||
---|---|---|
478 | Yeah, I haven't rebase against change that introduced maxLvlRank yet. | |
479 | No, it is not redundant. The lvlTypes current is a mapping for (tid, loopid) => dlt, not (tid, level) => dlt. I think storing a pair makes more sense than introducing a complete (tid, level) => dlt map here because the dlt is only required here when there are non-trivial index expressions on the level. | |
mlir/lib/Dialect/SparseTensor/Transforms/LoopEmitter.cpp | ||
43 | Sry, probably it get overlooked during rebasing | |
mlir/lib/Dialect/SparseTensor/Transforms/LoopEmitter.h | ||
229–233 | I agree with you on this, in fact, see my comment at L222, I will do it in a separate patch though. | |
250 | I will change the comment, it means whether it is the initial tensor that has not yet been sliced. | |
386–388 | Yeah, I am writing a paper on this (but still at very early stage), I will share it with you later when it is more or less complete. |
rebase + address comments.
mlir/lib/Dialect/SparseTensor/Transforms/LoopEmitter.h | ||
---|---|---|
246 | It should work, I will add a TODO here and submit the change in a separate patch. |
mlir/include/mlir/Dialect/SparseTensor/Utils/Merger.h | ||
---|---|---|
446 | As a TODO? | |
446 | Please mark such comments with a TODO, since you clearly have an idea on how to to it better. That way we can periodically grep for TODO's and fix them. Unless you think we will never do that, and then this is a note to self that should not be here. | |
476–477 | since you added a parameter, you also need to update this comment | |
481 | Top level comments usually apply to the block of code, I find e.g. assert(!loopToDependencies[i][t].has_value()); // must be first definition a lot more readable, since you follow it direclty with the make pair/pushback | |
492 | The non-trivial concept is used more widely now, but it would still be nice to define this per file, or at least per first occurrence what non-trivial really means. Or perhaps we should start using more standard terminology on affine expressions? | |
520 | tensor level with index expression on it, reads awkward how about must be a tensor level that contains a non-trivial index expression | |
mlir/lib/Dialect/SparseTensor/Transforms/LoopEmitter.cpp | ||
663 | If the slice | |
665 | I find this block of code extremely hard to read. | |
804 | A note "make sure" is very ambiguous. Is that a note to self, or something that the code actively does. | |
805 | appears first than normal tensors appears before normal tensors? | |
1194 | I agree the else is very long and deep Why not if (!resolved) { genSlice continue; } | |
1297 | Ok, this block is where all the magic happens ;-) | |
1768 | The next | |
1846 | Sets (or Increment), but use one style | |
mlir/lib/Dialect/SparseTensor/Transforms/LoopEmitter.h | ||
85 | comments with should or must are a bit dangling unless you say what happens when this assumption is not true | |
138 | Why did you change Exits -> exit? Original seems okay | |
209–210 | since we have several nested structs, can you give each a short comment (as documentation, and to improve readability( | |
228 | Here and below (and above), period at end | |
390 | const ref? | |
mlir/lib/Dialect/SparseTensor/Transforms/Sparsification.cpp | ||
1674 | I would start with the same comment as in the else, and state it in the affirmative rather than the speculative // End either a for-loop or a while-loop that iterates over a slice. |
simplify code.
mlir/lib/Dialect/SparseTensor/Transforms/LoopEmitter.cpp | ||
---|---|---|
665 | better now? |
mlir/include/mlir/Dialect/SparseTensor/Utils/Merger.h | ||
---|---|---|
481 | can we make this lvl < .... part also a helper method | |
mlir/lib/Dialect/SparseTensor/Transforms/LoopEmitter.cpp | ||
84 | This computation does not match my mental interpretation of the text above (L79) | |
88 | // offset adds very little, Either use a sentence or remove | |
254 | why the empty lines here? | |
319 | We need depends - 1 slices to make sure you don't read depends as part of the sentence | |
433 | The comment applies to the assert, but the declaration is in between | |
474 | Please elaborate. "Pop out" is not at all representative for what follows | |
665 | Yes, although it could still use a bit more doc on what each block does (on entry of each block). | |
750 | isn't that always the case here? Should that not be part of the method description then? | |
1300 | I think this still needs some work to make reading the block easier. The problem is that you have very concise comments in the header So I would still give every implementation function here an entry comment, but one that shows what is generated, using some pseudo-code of the output WDYT? | |
1313 | this one seems out of place (all others generate stuff) | |
1370 | here and a few other place, no period at end, please make once last pass over all new comments here | |
mlir/lib/Dialect/SparseTensor/Transforms/LoopEmitter.h | ||
85 | I think what is still missing is whether it is enforced (viz. asserts fail when trying to set it) So, something like Clients are responsible for ensuring that the order of the returned (I think my original comment was really on when I see "should" or "must", who is to blame in the end ;-) | |
256 | Wren can correct me if I am wrong, but I think this needs to be minimum, right (as in smallest value, and not the lowest according to some other measure)? | |
322 | is exceeds -> exceeds but more importantly, I would state this, We break out of the loop when the coordindate exceeds the slideSize. | |
386 | the most recent slice (singular) | |
394 | perhaps we should discuss somewhere else, but we use "unsigned" at most places, and size_t only for local operations, or inside casts and asserts | |
458 | as follows? | |
463 | period at end. "to allocate top level local" makes very little sense when read in isolation. Just say what code fragment this points to | |
mlir/lib/Dialect/SparseTensor/Transforms/Sparsification.cpp | ||
87 | I know this was already there, but can we use override here to make it more clear that we implementing the base visitor class? // Visitor method overrides. section? | |
mlir/lib/Dialect/SparseTensor/Utils/Merger.cpp | ||
397–399 | has the locate property as well |
mlir/include/mlir/Dialect/SparseTensor/Utils/Merger.h | ||
---|---|---|
481 | Why do you need this extra assertion? The isValidLevel assertion already ensures that the lvl is valid for the tensor t. Therefore, rather than checking the lvl twice, the rest of the code should instead maintain the invariant that levelToDependentLoop[t].size() == lvlTypes[t].size(). |
mlir/include/mlir/Dialect/SparseTensor/Utils/Merger.h | ||
---|---|---|
479 | There is a lot of unnecessary complexity and redundancy in storing all of:
As I mentioned in my earlier comment, we can easily reconstruct the desired (tid, lvl) -> dlt map via [](t, l) { auto i = lvlToLoop[t][l]; return i ? lvlTypes[t][*i] : Undef; }. Therefore, we always have that loopToDependencies[i][t] == make_pair(l, reconstructedLvlTypes[t][l]). Consequently: since the first part of loopToDependencies has that every (t,i) pair determines l, it is trivial to construct the required (t,l) pair for passing to reconstructedLvlTypes; and since it's trivial to define reconstructedLvlTypes, therefore there is no benefit to storing this redundant information. And as I said before, whenever we store redundant information that means we must also therefore take pains to ensure that all the copies of that information remain consistent. I agree that it would be nice to store the (tid, lvl) -> dlt map directly, and to use that in lieu of the current (tid, loopid) -> dlt map. Especially since the former can be quickly constructed from the types of the tensors, and doesn't require knowing anything about lvlToLoop/loopToLvl. However, regardless of which one we store, the point remains the same: there's no benefit to loopToDependencies storing this information redundantly, and if it stores redundant information anyways then it needs to ensure that it remains consistent with the (tid, {lvl,loopid}) -> dlt map. |
address comments.
mlir/include/mlir/Dialect/SparseTensor/Utils/Merger.h | ||
---|---|---|
481 | I found this is actually a redundant check, if the lvl is valid then it is definitely inbound. | |
mlir/lib/Dialect/SparseTensor/Transforms/Sparsification.cpp | ||
87 | I added a comment, this is non-vritual function, so I did not use override here. |
mlir/include/mlir/Dialect/SparseTensor/Utils/Merger.h | ||
---|---|---|
479 | I agree that we can probably clean it up, but I will address this in separate patches through ;-) |
mlir/lib/Dialect/SparseTensor/Transforms/LoopEmitter.cpp | ||
---|---|---|
1553 | I first though this was commented out code ;-) So make it Generate: code |
Hi @Peiming, the buildbots are failing (e.g. https://lab.llvm.org/buildbot/#/builders/160/builds/19165) - could you please fix it?
Yeah, I saw it. but the warning seems to be unrelated to this change... I will take a look
@vzakhari this is not the patch that triggers the complaining. If you are going to revert, please make sure you revert the right one, which is https://reviews.llvm.org/D148565
@vzakhari I do not think my patch caused the error, see https://lab.llvm.org/buildbot/#/builders/160/builds/19161, there was already the same warning (but I do not know why it was not treated as errors).
let me explain what I see: I looked at https://lab.llvm.org/buildbot/#/builders/160 and the first failing build was 19162: https://lab.llvm.org/buildbot/#/builders/160/builds/19162; it points to D148565. The build issue is shown in stdio section:
FAILED: tools/mlir/lib/Dialect/SparseTensor/Transforms/CMakeFiles/obj.MLIRSparseTensorTransforms.dir/SparseGPUCodegen.cpp.o /usr/local/bin/c++ -DGTEST_HAS_RTTI=0 -DMLIR_CUDA_CONVERSIONS_ENABLED=0 -DMLIR_ROCM_CONVERSIONS_ENABLED=0 -D_DEBUG -D_GLIBCXX_ASSERTIONS -D_GNU_SOURCE -D_LIBCPP_ENABLE_ASSERTIONS -D__STDC_CONSTANT_MACROS -D__STDC_FORMAT_MACROS -D__STDC_LIMIT_MACROS -I/home/tcwg-buildbot/worker/flang-aarch64-latest-gcc/build/tools/mlir/lib/Dialect/SparseTensor/Transforms -I/home/tcwg-buildbot/worker/flang-aarch64-latest-gcc/llvm-project/mlir/lib/Dialect/SparseTensor/Transforms -I/home/tcwg-buildbot/worker/flang-aarch64-latest-gcc/build/include -I/home/tcwg-buildbot/worker/flang-aarch64-latest-gcc/llvm-project/llvm/include -I/home/tcwg-buildbot/worker/flang-aarch64-latest-gcc/llvm-project/mlir/include -I/home/tcwg-buildbot/worker/flang-aarch64-latest-gcc/build/tools/mlir/include -fPIC -fno-semantic-interposition -fvisibility-inlines-hidden -Werror=date-time -Wall -Wextra -Wno-unused-parameter -Wwrite-strings -Wcast-qual -Wno-missing-field-initializers -pedantic -Wno-long-long -Wimplicit-fallthrough -Wno-maybe-uninitialized -Wno-class-memaccess -Wno-redundant-move -Wno-pessimizing-move -Wno-noexcept-type -Wdelete-non-virtual-dtor -Wsuggest-override -Wno-comment -Wno-misleading-indentation -Wctad-maybe-unsupported -fdiagnostics-color -ffunction-sections -fdata-sections -O3 -DNDEBUG -fno-exceptions -funwind-tables -fno-rtti -UNDEBUG -std=c++17 -MD -MT tools/mlir/lib/Dialect/SparseTensor/Transforms/CMakeFiles/obj.MLIRSparseTensorTransforms.dir/SparseGPUCodegen.cpp.o -MF tools/mlir/lib/Dialect/SparseTensor/Transforms/CMakeFiles/obj.MLIRSparseTensorTransforms.dir/SparseGPUCodegen.cpp.o.d -o tools/mlir/lib/Dialect/SparseTensor/Transforms/CMakeFiles/obj.MLIRSparseTensorTransforms.dir/SparseGPUCodegen.cpp.o -c /home/tcwg-buildbot/worker/flang-aarch64-latest-gcc/llvm-project/mlir/lib/Dialect/SparseTensor/Transforms/SparseGPUCodegen.cpp In file included from ../llvm-project/mlir/lib/Dialect/SparseTensor/Transforms/SparseGPUCodegen.cpp:17: ../llvm-project/mlir/lib/Dialect/SparseTensor/Transforms/LoopEmitter.h: In member function ‘constexpr mlir::sparse_tensor::TensorLevel mlir::sparse_tensor::LoopEmitter::makeTensorLevel(mlir::sparse_tensor::TensorId, mlir::sparse_tensor::Level) const’: ../llvm-project/mlir/lib/Dialect/SparseTensor/Transforms/LoopEmitter.h:199:29: error: call to non-‘constexpr’ function ‘unsigned int mlir::sparse_tensor::LoopEmitter::getNumTensors() const’ 199 | return l * getNumTensors() + t; | ~~~~~~~~~~~~~^~ ../llvm-project/mlir/lib/Dialect/SparseTensor/Transforms/LoopEmitter.h:195:12: note: ‘unsigned int mlir::sparse_tensor::LoopEmitter::getNumTensors() const’ declared here 195 | unsigned getNumTensors() const { return tensors.size(); } | ^~~~~~~~~~~~~ 90.309 [1754/1/4349] Linking CXX shared library lib/libclang-cpp.so.17git ninja: build stopped: subcommand failed.
So it does point to the change from D148565.
As a TODO?