Skip to content

Feature Requests

298 results found

  1. Test Filter Builder: Enable filtering by Project Name via regular expressions

    Test Filter Builder already supports filtering by Full Name with Regular Expression, and it supports filtering by Project Name, but it doesn't appear that there's a way to filter by project name using wildcards/regex.

    I've got test projects containing pure unit tests, named .UnitTests as a convention, but the test classes' namespaces are the same as their SUT (system under test) class (the namespaces don't have ".UnitTests" in them).

    I'd like to trigger my (pure, fast) UnitTests to run when impacted, but not my integration tests (they're in separate projects named *.IntegrationTests).

    1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    0 comments  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  2. Allow user to set color of code coverage marks while a test is executing.

    Currently it only goes a few shades light and it's difficult to see if this is in progress or if it's already passed/failed. Being able to set this color to something like orange gives a clear indication that the test is in progress.

    Changing the graphic would also work, instead of a green circle it could be a green question mark.

    2 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  3. Add median, quartil, percentile or truncated mean to "Marker performance aggregation type"

    First run takes offen more time as following ones. Because of small count of runs in typical unittest scenario, the first outlier has too much weight. In this Case median, quartil, percentile or truncated mean are robust.

    Ps. It should be simple to implement this. )

    2 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  4. Deep code coverage tool

    I was thinking about a better code coverage tool. A tool that take longer to run, but show the real thing.

    This deep code coverage tool, should test each line of code. If a line is commented, at least one test should fail or an error should be outlined. The tool can do the check and report the lines that are not well covered.

    The result it's easier to remove the line or improve our tests for a better code coverage.

    1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    0 comments  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  5. ncrunch "absorb" solution folder structure to organize the tests

    Can we have ncrunch "absorb" solution folder structure to organise the tests in the test overview window? We currently have two folders (slow and fast) and resharper adheres to that struture (very convenient).

    I'd like ncrunch to do the same. Could you add that to the Tree structure option in the NCrunch Tests window?

    6 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    1 comment  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  6. Custom Engine Modes - Dynamic project name regex by changed project name

    Hello,

    we have the following naming conversion:
    Company.ComponentA.Foo.dll
    Company.ComponentA.Foo_uTest.dll

    I want the relevant unit test dlls to be selected based on the changed product projects.

    As an addition the mapping should be flexible and support:
    1. Just the unit test target of the changed product target should be used
    2. All unit test targets of a changed component should be used.

    Idea:
    1. Selection Pattern for the changed project
    '^(\w+[.]\w+[.]).*$' // Captures "Company.ComponentA."

    Then a string.Format line to create the regex pattern for the project name selection:

    '^{0}._uTest$' which would lead to "Company.ComponentA.._uTest"

    => All unit tests of…

    1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    0 comments  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  7. Custom Engine Modes - Project name matches regex

    Hello,
    I want to use regex to match project names because in our projects there is a naming rule:

    <Product-Assembly> has a corresponding unit test target with the name:
    <Product-Assembly>_uTest

    Therefore I want to define a rule like:
    Project name matches regex '^.*_uTest$'

    With best regards,
    stefan

    1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    0 comments  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  8. Allow mouse-click navigation on stack trace even when instrumentation is off

    When instrumentation is ON you can navigate from error stack trace to the code with a single mouse-click (very similar to ReSharper stack trace explorer).
    When instrumentation is OFF then mouse-click functionality stops working because NCrunch-specific paths are used in this case. ReSharper Stack Trace Explorer still works fine and is able to handle stack traces from NCrunch correctly.
    Original issue reports is here: http://forum.ncrunch.net/yaf_postst1513_Error-StackTrace-is-not-navigatable-when-instrumentation-is-off.aspx

    2 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  9. Add progress bar to Metrics tree view

    I would find the metrics tree view more useful if the code coverage % column had a progress bar in the background. The numbers are great of course, but a progress bar would be much easier to glance at, especially when looking at a large number of rows.

    DotCover has a view that shows what I mean well: https://www.jetbrains.com/dotcover/features/img/dotcover-app.png

    3 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  10. Add an attribute like [Isolated] which applies to each member of a fixture rather than the fixture as a whole

    The [Isolated] attribute currently only provides isolation for the fixture, not each test within it. Where tests need to be isolated, they need to be individually annotated.

    It would be nice to have an attribute like [IsolateEachMember] which could apply to each test in a fixture.

    4 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  11. Install the service as "Automatic (Delayed Start)"

    Really small change, but can the Grid Node server install with the service option of Delayed Start rather than Automatic.

    The only difference is that there will be roughly a 2 minute delay before NCrunch gets started after a server restart, but it means that the server will have had chance to settle down before NCrunch starts running builds and tests.

    2 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  12. Test result icon next to method name

    In the text editor when all the test methods are collapsed, we cannot see which test have failed.

    If an icon next to the method declaration was also displayed we could know which one needs to be expanded and work on without the need to look at the overall NCrunch test window.

    4 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    1 comment  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  13. Implement Ctrl+A shortcut in NCrunch configuration grid

    For the moment it's possible to select more than one project but only holding Ctrl key and clicking on each project.

    3 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  14. Separate 'successfully completed' and 'failed tasks' in the Processing Queue window

    The NCrunch Processing Queue window currently has 3 toggles:
    - Pending tasks
    - Tasks being processed
    - Completed tasks

    The last of these has both the successfully completed and the failed tasks. However, I find that confusing and inconsistent:

    • The icon on the toggle button is a checkmark. The icon for successful tasks is a checkmark, while the icon for a failed task is a red "X".
    • The NCrunch Tests window also has successful and failed "tasks", in this case the tests. However here there are two toggles, one for successful and one for failed tests.
    3 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    1 comment  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  15. Support Windows Runtime Components

    Support C++ Windows Runtime Components. These are used in Windows Store projects. One of the cool things about WinRT is that you can mix and match languages. I'm not talking about running tests against C++, just allow referencing them. Currently even referencing a component breaks NCrunch compilation for the whole solution.

    4 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  16. Add filter in "test popup" window

    So I navigate code and possible spot an issue, I click on line to bring up the "tests covering this line" and I am interested in a test containing a string e.g. "NoOfQueries" but I am a bit overwelmed with the no of tests for this code so I cannot find it.

    Current workaround is to pin all tests covering it, and then add a text filter in the Tests window (and then unpin all tests which kindof ruins my current "pinned tests scope").

    1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    0 comments  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  17. cover

    When collapsing a method have another icon next to it indicating the coverage.

    It can have 5 colors:
    Black - no cover
    Green - 100% covering tests. All tests pass
    Red - 100% covering tests with at least one failing test
    Blue - Contains some passing tests but also there are lines of code that are not covered by tests
    Orange - Contains some failing tests but also there are lines of code that are not covered by tests

    2 votes

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  18. Allow a test project to be excluded from creating inline code coverage

    Inline code coverage is great for unit tests, but when integration tests are also instrumented we lose the granularity of coverage.

    N.B Our integration tests are for testing the interaction of our classes either with each other, or 3rd party libraries. So running them in ncrunch and getting test failure notifications straight away is useful for us.

    1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    1 comment  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  19. 1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    0 comments  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  20. 1 vote

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)

    We’ll send you updates on this idea

    0 comments  ·  Admin →
    How important is this to you?

    We're glad you're here

    Please sign in to leave feedback

    Signed in as (Sign out)
  • Don't see your idea?

Feature Requests

Categories

Feedback and Knowledge Base