[yt-dev] Testing Intervention

Casey W. Stark caseywstark at gmail.com
Tue Sep 25 16:12:09 PDT 2012


Hey Sam.

I was suggesting...

yt-hg/
  test/
  yt/

So that it is separate from the yt package.

- Casey


On Tue, Sep 25, 2012 at 4:05 PM, Sam Skillman <samskillman at gmail.com> wrote:

> Would root in our case be yt-hg/yt, in which case the tests would live at
> yt-hg/yt/tests/ ?
>
> Sorry if I misunderstood.
> Sam
>
>
> On Tue, Sep 25, 2012 at 5:00 PM, Anthony Scopatz <scopatz at gmail.com>wrote:
>
>> Hi Casey,
>>
>> The reason I suggested that we do it this way originally was because this
>> is the way that most other packages in the scientific python ecosystem are
>> set up.  If you look at numpy, scipy, cython, pytables, ipython, etc. they
>> all have the following file system structure:
>>
>> root/
>>   module1.py
>>   module2.py
>>   tests/
>>     test_module2.py
>>     test_module1.py
>>   subpack1/
>>     module3.py
>>     tests/
>>       test_module3.py
>>
>> I think that it is done this way to make it easy/possible for UnitTest
>> and nose to autodetect the tests.  I am not saying that we have to follow
>> this de facto standard.  However, I don't really see a reason why we
>> shouldn't either.  If you do, please let me know ;)
>>
>> Notably, matplotlib does not follow this standard.  But then again, since
>> they are testing GUI and rendering code their tests are a level of
>> complexity above what we need.
>>
>> Be Well
>> Anthony
>>
>> On Tue, Sep 25, 2012 at 5:50 PM, Casey W. Stark <caseywstark at gmail.com>wrote:
>>
>>> Hey Matt, Anthony.
>>>
>>> Looks like a good start to nosetests. One small thing -- would anyone be
>>> opposed to moving these out of the yt directory and into a root "test"
>>> directory?
>>>
>>> I can take on CIC deposition and interpolation wherever those are.
>>>
>>> - Casey
>>>
>>>
>>> On Tue, Sep 25, 2012 at 11:20 AM, Matthew Turk <matthewturk at gmail.com>wrote:
>>>
>>>> Hi all,
>>>>
>>>> As a note, Anthony just added this to the 3.0 base.  Here's a set of
>>>> tests:
>>>>
>>>>
>>>> https://bitbucket.org/yt_analysis/yt-3.0/src/52a1b2ecea94/yt/utilities/lib/tests/test_alt_ray_tracers.py
>>>>
>>>> and here's a utilities section he added:
>>>>
>>>> https://bitbucket.org/yt_analysis/yt-3.0/src/52a1b2ecea94/yt/testing.py
>>>>
>>>> -Matt
>>>>
>>>> On Mon, Sep 24, 2012 at 10:01 PM, Matthew Turk <matthewturk at gmail.com>
>>>> wrote:
>>>> > Hey Casey,
>>>> >
>>>> > On Mon, Sep 24, 2012 at 5:39 PM, Casey W. Stark <
>>>> caseywstark at gmail.com> wrote:
>>>> >> Hi Matt.
>>>> >>
>>>> >> Glad my example was useful in some way. I guess knowing exactly
>>>> which Cython
>>>> >> routines to test for what is what I meant about where to start.
>>>> >
>>>> > I completely understand -- and that's a failure of my part in the
>>>> > code.  So I think what might help is if we start looking at the
>>>> > routines that could be tested, and then I can jump in when it looks
>>>> > like they're buried.  I will also volunteer to try to refactor this
>>>> > code, once it's tested, to make it clearer what is where and why.
>>>> >
>>>> >>
>>>> >> Thanks for the tip about the stream frontend.
>>>> >
>>>> > No prob, it should be documented better.
>>>> >
>>>> > In terms of generating data, here's an example of what I was meaning:
>>>> >
>>>> > https://hub.yt-project.org/nb/ompxtg
>>>> >
>>>> > This sets up a random (but small-valued) background density and then
>>>> > applies a couple top hat spheres on top of it.  I think this could be
>>>> > a good starting point for unit testing.  I looked at the code that
>>>> > exists for flagging cells and refining them, and it's currently a bit
>>>> > too specific for this, as it was designed to take RAMSES data and
>>>> > partition it, which operates slightly differently.  I'll take a pass
>>>> > at extracting it and making it work in this case.
>>>> >
>>>> > The way I see it working is that one would set up the operators, as is
>>>> > done here, and then a progressive routine would be applied, something
>>>> > like this:
>>>> >
>>>> > while grids.flag() > 0:
>>>> >     grids.refine()
>>>> >     for operator in operators: operator.apply(grids)
>>>> > pf = grids.convert_to_dataset()
>>>> >
>>>> > The missing step in the code base is mostly just the refinement, as
>>>> > for the smoothed covering grids we have the necessary machinery to
>>>> > interpolate from one level to the next.  Adding more types of
>>>> > operators to this library would be beneficial as well.  The
>>>> > finalization step at the end would then convert the collection of
>>>> > grids into a Stream dataset.  So with only a couple lines of code we
>>>> > could in all likelihood be able to generate in-memory datasets.
>>>> >
>>>> > (This underscores in my mind why it would be awesome to have GDF in
>>>> > more places, as with this kind of machinery we've just written initial
>>>> > conditions generation, especially since we have a GDF writer
>>>> > already...)
>>>> >
>>>> > Anyway, what might be really helpful is if interested people would
>>>> > volunteer to address a few of the testable areas?  Then we can start
>>>> > pushing and identifying problem areas.  I'll volunteer to handle at
>>>> > least the data container stuff, and whatever else slips through the
>>>> > cracks; although my time will be somewhat limited in the very near
>>>> > future, I will try to make this a priority.
>>>> >
>>>> > -Matt
>>>> >
>>>> >>
>>>> >> - Casey
>>>> >>
>>>> >>
>>>> >> On Mon, Sep 24, 2012 at 1:52 PM, Matthew Turk <matthewturk at gmail.com>
>>>> wrote:
>>>> >>>
>>>> >>> Hey Casey and Anthony,
>>>> >>>
>>>> >>> On Mon, Sep 24, 2012 at 4:20 PM, Casey W. Stark <
>>>> caseywstark at gmail.com>
>>>> >>> wrote:
>>>> >>> > Hi Anthony.
>>>> >>> >
>>>> >>> > I completely agree that we should target the level of functions
>>>> actually
>>>> >>> > performing the projection rather than yt's organization. The mock
>>>> >>> > frontend
>>>> >>> > suggestion was just a hack to get there. I don't know if there's
>>>> a way
>>>> >>> > around it though...
>>>> >>> >
>>>> >>> > Here's an example of what I sorted through to get to projections:
>>>> >>> > - Load a test plotfile, check pf.h.proj to find it's source.
>>>> >>> > - Read through data_objects/hierarchy.py and
>>>> >>> > utilities/parallel_tools/parallel_analysis_interface.py to find
>>>> where
>>>> >>> > proj
>>>> >>> > is attached, can't find it.
>>>> >>> > - The proj docstring says it is a reference to AMRQuadProj. Can't
>>>> find a
>>>> >>> > class by that name.
>>>> >>> > - Search data_objects sources for "proj", find AMRProjBase.
>>>> >>> >
>>>> >>> > So it looks like the functionality is wrapped up in the
>>>> __project_level
>>>> >>> > and
>>>> >>> > _project_grid methods. I can't think of a way to test those
>>>> without
>>>> >>> > creating
>>>> >>> > an AMRProjBase, and that requires a staticoutput object.
>>>> >>>
>>>> >>> You're right, the projection stuff as *projections* is not easy to
>>>> >>> test.  But in terms of testing the underlying code, which is wrapped
>>>> >>> up in a Cython class called QuadTree, I think it could be done.  The
>>>> >>> steps you're describing are actually all part of the existing answer
>>>> >>> testing machinery, which performs a couple things and verifies that
>>>> >>> they don't change over time:
>>>> >>>
>>>> >>> 1) Project some fields from the disk
>>>> >>> 2) Project a couple derived fields
>>>> >>> 3) Project a derived field that requires spatial derivatives
>>>> >>> 4) Project the "Ones" field, which should be 1.0 everywhere.
>>>> >>>
>>>> >>> So these things are done, but it is also possible that the specific
>>>> >>> quadtree functionality could be tested, in isolation from the
>>>> >>> projection.  I think this may be oneo f the things Anthony is
>>>> talking
>>>> >>> about -- answer testing can handle the big, complex items, and by
>>>> >>> breaking down to the fundamentals we can address isolated items
>>>> from a
>>>> >>> unit testing perspective.
>>>> >>>
>>>> >>> >
>>>> >>> > So unfortunately, I think it would still come down to having a
>>>> fake
>>>> >>> > frontend. It's not ideal, but it seems like any more isolation
>>>> would
>>>> >>> > require
>>>> >>> > big rewrites to yt.
>>>> >>>
>>>> >>> One fun thing that is not usually known is that we have a fake
>>>> >>> frontend already, it just doesn't get used much.  It's called the
>>>> >>> "Stream" frontend and it was designed originally to be used in
>>>> >>> ParaView, but now gets used by the (new,
>>>> not-yet-documented/released)
>>>> >>> load_uniform_grid function as well as by Hyperion, the RT code by
>>>> Tom
>>>> >>> R.  It can set up AMR as well as static mesh.  It's not terribly
>>>> well
>>>> >>> documented, but there are examples on the wiki.
>>>> >>>
>>>> >>> One thing I've been thinking about is actually creating a couple
>>>> fake
>>>> >>> outputs, which could be defined analytically with spheres of
>>>> >>> overdensity inside them.  In principle, if we added refinement
>>>> >>> criteria, we could make this relatively complex data that was
>>>> defined
>>>> >>> with only a few lines of code, but spun up a big in-memory dataset.
>>>> >>>
>>>> >>> (This exact thing is on my list of things to do and then to output
>>>> in
>>>> >>> GDF, by the way...)
>>>> >>>
>>>> >>> That I think could come, down the road a bit.  The refinement
>>>> criteria
>>>> >>> wouldn't be too bad to implement, especially since we already have
>>>> the
>>>> >>> grid splitting routines.  I just don't think we should focus on it
>>>> at
>>>> >>> the moment.  But the uniform grid creation and loading works already
>>>> >>> -- I used it this morning.  You can do it with:
>>>> >>>
>>>> >>> from yt.frontends.stream.api import load_uniform_grid
>>>> >>> ug = load_uniform_grid({"VelocityNorm":data1, "Density":data2},
>>>> [359,
>>>> >>> 359, 359], 1.0)
>>>> >>>
>>>> >>> the list is the dimensions of the data and the value is the to-cm
>>>> >>> conversion.
>>>> >>>
>>>> >>> >
>>>> >>> > Of course, I could be missing something. Matt, can you think of a
>>>> better
>>>> >>> > way?
>>>> >>>
>>>> >>> I think for this specific example (and your damningly complex
>>>> tracing
>>>> >>> of things through the source ...) the easiest thing to do is isolate
>>>> >>> the Cython routine, which it seems I was able to do only because I
>>>> >>> wrote it and which seems quite buried in the code, and to also
>>>> provide
>>>> >>> high-level machinery for faking a frontend.
>>>> >>>
>>>> >>> -Matt
>>>> >>>
>>>> >>> >
>>>> >>> > - Casey
>>>> >>> >
>>>> >>> >
>>>> >>> > On Mon, Sep 24, 2012 at 11:02 AM, Anthony Scopatz <
>>>> scopatz at gmail.com>
>>>> >>> > wrote:
>>>> >>> >>
>>>> >>> >> Helo Casey,
>>>> >>> >>
>>>> >>> >> Sorry for taking the whole weekend to respond.
>>>> >>> >>
>>>> >>> >>>> I would like to help with this, but it's difficult to figure
>>>> out
>>>> >>> >>>> where
>>>> >>> >>>> to start.
>>>> >>> >>
>>>> >>> >>
>>>> >>> >> Not to worry. I think that any of the items listed at the bottom
>>>> of
>>>> >>> >> Matt's
>>>> >>> >> original email
>>>> >>> >> would be a great place to start.
>>>> >>> >>
>>>> >>> >>>>
>>>> >>> >>>>
>>>> >>> >>>> Say I want to test projections. I make a fake 3D density
>>>> field, maybe
>>>> >>> >>>> something as simple as np.arange(4**3).reshape((4, 4, 4)). I
>>>> write
>>>> >>> >>>> down the
>>>> >>> >>>> answer to the x-projection. Now all I need to do is call
>>>> >>> >>>> assert_allclose(yt_result, answer, rtol=1e-15), but I don't
>>>> know what
>>>> >>> >>>> pieces
>>>> >>> >>>> of low-level yt stuff to call to get to `yt_result`. Hopefully
>>>> that's
>>>> >>> >>>> clear...
>>>> >>> >>>>
>>>> >>> >>>> Maybe this comes down to creating a fake frontend we can attach
>>>> >>> >>>> fields
>>>> >>> >>>> to?
>>>> >>> >>
>>>> >>> >>
>>>> >>> >> Actually, I disagree with this strategy, as I told Matt when we
>>>> spoke
>>>> >>> >> last
>>>> >>> >> week.
>>>> >>> >> What is important is that we test the science and math parts of
>>>> the
>>>> >>> >> code
>>>> >>> >> before, if ever, dealing with the software architecture that
>>>> surrounds
>>>> >>> >> them.
>>>> >>> >>
>>>> >>> >> Let's taking your example of projections.  What we need to test
>>>> is the
>>>> >>> >> actual function
>>>> >>> >> or method which actually slogs through the projection
>>>> calculation.  In
>>>> >>> >> many cases in
>>>> >>> >> yt these functions are not directly attached to the front end
>>>> but live
>>>> >>> >> in
>>>> >>> >> analysis, visualization
>>>> >>> >> or utilities subpackages.   It is these such packages that we
>>>> should
>>>> >>> >> worry
>>>> >>> >> about testing.
>>>> >>> >> We can easily create routines to feed them sample data.
>>>> >>> >>
>>>> >>> >> On the other hand, testing or mocking things like frontends
>>>> should be a
>>>> >>> >> very low priority.
>>>> >>> >> At the end of the day what you are testing here is pulling in
>>>> data from
>>>> >>> >> disk or other
>>>> >>> >> sources.  Effectively, this is just re-testing functionality
>>>> present in
>>>> >>> >> h5py, etc.  That is not
>>>> >>> >> really our job.  Yes, in a perfect world, front ends would be
>>>> tested
>>>> >>> >> too.
>>>> >>> >> But I think that the
>>>> >>> >> priority should be placed on things like the KDTree.
>>>> >>> >>
>>>> >>> >> Be Well
>>>> >>> >> Anthony
>>>> >>> >>
>>>> >>> >>>>
>>>> >>> >>>>
>>>> >>> >>>> - Casey
>>>> >>> >>>>
>>>> >>> >>>>
>>>> >>> >>>> On Fri, Sep 21, 2012 at 2:42 PM, Matthew Turk <
>>>> matthewturk at gmail.com>
>>>> >>> >>>> wrote:
>>>> >>> >>>>>
>>>> >>> >>>>> Hi all,
>>>> >>> >>>>>
>>>> >>> >>>>> As some of you have seen (at least Stephen), I filed a ticket
>>>> this
>>>> >>> >>>>> morning about increasing testing coverage.  The other night
>>>> Anthony
>>>> >>> >>>>> and I met up in NYC and he had something of an "intervention"
>>>> about
>>>> >>> >>>>> the sufficiency of answer testing for yt; it didn't take too
>>>> much
>>>> >>> >>>>> work
>>>> >>> >>>>> on his part to convince me that we should be testing not just
>>>> >>> >>>>> against
>>>> >>> >>>>> a gold standard, but also performing unit tests.  In the past
>>>> I had
>>>> >>> >>>>> eschewed unit testing simply because the task of mocking data
>>>> was
>>>> >>> >>>>> quite tricky, and by adding tests that use smaller bits we
>>>> could
>>>> >>> >>>>> cover
>>>> >>> >>>>> unit testable areas with answer testing.
>>>> >>> >>>>>
>>>> >>> >>>>> But, this isn't really a good strategy.  Let's move to having
>>>> both.
>>>> >>> >>>>> The testing infrastructure he recommends is the
>>>> nearly-omnipresent
>>>> >>> >>>>> nose:
>>>> >>> >>>>>
>>>> >>> >>>>> http://nose.readthedocs.org/en/latest/
>>>> >>> >>>>>
>>>> >>> >>>>> The ticket to track this is here:
>>>> >>> >>>>>
>>>> >>> >>>>>
>>>> >>> >>>>>
>>>> >>> >>>>>
>>>> https://bitbucket.org/yt_analysis/yt/issue/426/increase-unit-test-coverage
>>>> >>> >>>>>
>>>> >>> >>>>> There are a couple sub-items here:
>>>> >>> >>>>>
>>>> >>> >>>>> 1) NumPy's nose test plugins provide a lot of necessary
>>>> >>> >>>>> functionality
>>>> >>> >>>>> that we have reimplemented in the answer testing utilities.
>>>>  I'd
>>>> >>> >>>>> like
>>>> >>> >>>>> to start using the numpy plugins, which include things like
>>>> >>> >>>>> conditional test execution, array comparisons, "slow" tests,
>>>> etc
>>>> >>> >>>>> etc.
>>>> >>> >>>>> 2) We can evaluate, using conditional test execution, moving
>>>> to nose
>>>> >>> >>>>> for answer testing.  But that's not on the agenda now.
>>>> >>> >>>>> 3) Writing tests for nose is super easy, and running them is
>>>> too.
>>>> >>> >>>>> Just
>>>> >>> >>>>> do:
>>>> >>> >>>>>
>>>> >>> >>>>> nosetest -w yt/
>>>> >>> >>>>>
>>>> >>> >>>>> when in your source directory.
>>>> >>> >>>>>
>>>> >>> >>>>> 4) I've written a simple sample here:
>>>> >>> >>>>>
>>>> >>> >>>>>
>>>> >>> >>>>>
>>>> >>> >>>>>
>>>> https://bitbucket.org/yt_analysis/yt-3.0/src/da10ffc17f6d/yt/utilities/tests/test_interpolators.py
>>>> >>> >>>>>
>>>> >>> >>>>> 5) I'll handle writing up some mock data that doesn't require
>>>> >>> >>>>> shipping
>>>> >>> >>>>> lots of binary files, which can then be used for checking
>>>> things
>>>> >>> >>>>> that
>>>> >>> >>>>> absolutely require hierarchies.
>>>> >>> >>>>>
>>>> >>> >>>>> --
>>>> >>> >>>>>
>>>> >>> >>>>> The way to organize tests is easy.  Inside each directory with
>>>> >>> >>>>> testable items create a new directory called "tests", and in
>>>> here
>>>> >>> >>>>> toss
>>>> >>> >>>>> some scripts.  You can stick a bunch of functions in those
>>>> scripts.
>>>> >>> >>>>>
>>>> >>> >>>>> Anyway, I'm going to start writing more of these (in the main
>>>> yt
>>>> >>> >>>>> repo,
>>>> >>> >>>>> and this change will be grafted there as well) and I'll write
>>>> back
>>>> >>> >>>>> once the data mocking is ready.  I'd like it if we started
>>>> >>> >>>>> encouraging
>>>> >>> >>>>> or even mandating simple tests (and/or answer tests) for
>>>> >>> >>>>> functionality
>>>> >>> >>>>> that gets added, but that's a discussion that should be held
>>>> >>> >>>>> separately.
>>>> >>> >>>>>
>>>> >>> >>>>> The items on the ticket:
>>>> >>> >>>>>
>>>> >>> >>>>>  * kD-tree for nearest neighbor
>>>> >>> >>>>>  * Geometric selection routines
>>>> >>> >>>>>  * Profiles
>>>> >>> >>>>>  * Projections -- underlying quadtree
>>>> >>> >>>>>  * Data object selection of data containers
>>>> >>> >>>>>  * Data object selection of points
>>>> >>> >>>>>  * Orientation class
>>>> >>> >>>>>  * Pixelization
>>>> >>> >>>>>  * Color maps
>>>> >>> >>>>>  * PNG writing
>>>> >>> >>>>>
>>>> >>> >>>>> Is anyone willing to claim any additional items that they
>>>> will help
>>>> >>> >>>>> write unit tests for?
>>>> >>> >>>>>
>>>> >>> >>>>> -Matt
>>>> >>> >>>>> _______________________________________________
>>>> >>> >>>>> yt-dev mailing list
>>>> >>> >>>>> yt-dev at lists.spacepope.org
>>>> >>> >>>>> http://lists.spacepope.org/listinfo.cgi/yt-dev-spacepope.org
>>>> >>> >>>>
>>>> >>> >>>>
>>>> >>> >>>>
>>>> >>> >>>> _______________________________________________
>>>> >>> >>>> yt-dev mailing list
>>>> >>> >>>> yt-dev at lists.spacepope.org
>>>> >>> >>>> http://lists.spacepope.org/listinfo.cgi/yt-dev-spacepope.org
>>>> >>> >>>>
>>>> >>> >>
>>>> >>> >>
>>>> >>> >> _______________________________________________
>>>> >>> >> yt-dev mailing list
>>>> >>> >> yt-dev at lists.spacepope.org
>>>> >>> >> http://lists.spacepope.org/listinfo.cgi/yt-dev-spacepope.org
>>>> >>> >>
>>>> >>> >
>>>> >>> >
>>>> >>> > _______________________________________________
>>>> >>> > yt-dev mailing list
>>>> >>> > yt-dev at lists.spacepope.org
>>>> >>> > http://lists.spacepope.org/listinfo.cgi/yt-dev-spacepope.org
>>>> >>> >
>>>> >>> _______________________________________________
>>>> >>> yt-dev mailing list
>>>> >>> yt-dev at lists.spacepope.org
>>>> >>> http://lists.spacepope.org/listinfo.cgi/yt-dev-spacepope.org
>>>> >>
>>>> >>
>>>> >>
>>>> >> _______________________________________________
>>>> >> yt-dev mailing list
>>>> >> yt-dev at lists.spacepope.org
>>>> >> http://lists.spacepope.org/listinfo.cgi/yt-dev-spacepope.org
>>>> >>
>>>> _______________________________________________
>>>> yt-dev mailing list
>>>> yt-dev at lists.spacepope.org
>>>> http://lists.spacepope.org/listinfo.cgi/yt-dev-spacepope.org
>>>>
>>>
>>>
>>> _______________________________________________
>>> yt-dev mailing list
>>> yt-dev at lists.spacepope.org
>>> http://lists.spacepope.org/listinfo.cgi/yt-dev-spacepope.org
>>>
>>>
>>
>> _______________________________________________
>> yt-dev mailing list
>> yt-dev at lists.spacepope.org
>> http://lists.spacepope.org/listinfo.cgi/yt-dev-spacepope.org
>>
>>
>
> _______________________________________________
> yt-dev mailing list
> yt-dev at lists.spacepope.org
> http://lists.spacepope.org/listinfo.cgi/yt-dev-spacepope.org
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.spacepope.org/pipermail/yt-dev-spacepope.org/attachments/20120925/aef705ac/attachment.htm>


More information about the yt-dev mailing list