[yt-dev] Testing Intervention
Matthew Turk
matthewturk at gmail.com
Tue Sep 25 11:20:26 PDT 2012
Hi all,
As a note, Anthony just added this to the 3.0 base. Here's a set of tests:
https://bitbucket.org/yt_analysis/yt-3.0/src/52a1b2ecea94/yt/utilities/lib/tests/test_alt_ray_tracers.py
and here's a utilities section he added:
https://bitbucket.org/yt_analysis/yt-3.0/src/52a1b2ecea94/yt/testing.py
-Matt
On Mon, Sep 24, 2012 at 10:01 PM, Matthew Turk <matthewturk at gmail.com> wrote:
> Hey Casey,
>
> On Mon, Sep 24, 2012 at 5:39 PM, Casey W. Stark <caseywstark at gmail.com> wrote:
>> Hi Matt.
>>
>> Glad my example was useful in some way. I guess knowing exactly which Cython
>> routines to test for what is what I meant about where to start.
>
> I completely understand -- and that's a failure of my part in the
> code. So I think what might help is if we start looking at the
> routines that could be tested, and then I can jump in when it looks
> like they're buried. I will also volunteer to try to refactor this
> code, once it's tested, to make it clearer what is where and why.
>
>>
>> Thanks for the tip about the stream frontend.
>
> No prob, it should be documented better.
>
> In terms of generating data, here's an example of what I was meaning:
>
> https://hub.yt-project.org/nb/ompxtg
>
> This sets up a random (but small-valued) background density and then
> applies a couple top hat spheres on top of it. I think this could be
> a good starting point for unit testing. I looked at the code that
> exists for flagging cells and refining them, and it's currently a bit
> too specific for this, as it was designed to take RAMSES data and
> partition it, which operates slightly differently. I'll take a pass
> at extracting it and making it work in this case.
>
> The way I see it working is that one would set up the operators, as is
> done here, and then a progressive routine would be applied, something
> like this:
>
> while grids.flag() > 0:
> grids.refine()
> for operator in operators: operator.apply(grids)
> pf = grids.convert_to_dataset()
>
> The missing step in the code base is mostly just the refinement, as
> for the smoothed covering grids we have the necessary machinery to
> interpolate from one level to the next. Adding more types of
> operators to this library would be beneficial as well. The
> finalization step at the end would then convert the collection of
> grids into a Stream dataset. So with only a couple lines of code we
> could in all likelihood be able to generate in-memory datasets.
>
> (This underscores in my mind why it would be awesome to have GDF in
> more places, as with this kind of machinery we've just written initial
> conditions generation, especially since we have a GDF writer
> already...)
>
> Anyway, what might be really helpful is if interested people would
> volunteer to address a few of the testable areas? Then we can start
> pushing and identifying problem areas. I'll volunteer to handle at
> least the data container stuff, and whatever else slips through the
> cracks; although my time will be somewhat limited in the very near
> future, I will try to make this a priority.
>
> -Matt
>
>>
>> - Casey
>>
>>
>> On Mon, Sep 24, 2012 at 1:52 PM, Matthew Turk <matthewturk at gmail.com> wrote:
>>>
>>> Hey Casey and Anthony,
>>>
>>> On Mon, Sep 24, 2012 at 4:20 PM, Casey W. Stark <caseywstark at gmail.com>
>>> wrote:
>>> > Hi Anthony.
>>> >
>>> > I completely agree that we should target the level of functions actually
>>> > performing the projection rather than yt's organization. The mock
>>> > frontend
>>> > suggestion was just a hack to get there. I don't know if there's a way
>>> > around it though...
>>> >
>>> > Here's an example of what I sorted through to get to projections:
>>> > - Load a test plotfile, check pf.h.proj to find it's source.
>>> > - Read through data_objects/hierarchy.py and
>>> > utilities/parallel_tools/parallel_analysis_interface.py to find where
>>> > proj
>>> > is attached, can't find it.
>>> > - The proj docstring says it is a reference to AMRQuadProj. Can't find a
>>> > class by that name.
>>> > - Search data_objects sources for "proj", find AMRProjBase.
>>> >
>>> > So it looks like the functionality is wrapped up in the __project_level
>>> > and
>>> > _project_grid methods. I can't think of a way to test those without
>>> > creating
>>> > an AMRProjBase, and that requires a staticoutput object.
>>>
>>> You're right, the projection stuff as *projections* is not easy to
>>> test. But in terms of testing the underlying code, which is wrapped
>>> up in a Cython class called QuadTree, I think it could be done. The
>>> steps you're describing are actually all part of the existing answer
>>> testing machinery, which performs a couple things and verifies that
>>> they don't change over time:
>>>
>>> 1) Project some fields from the disk
>>> 2) Project a couple derived fields
>>> 3) Project a derived field that requires spatial derivatives
>>> 4) Project the "Ones" field, which should be 1.0 everywhere.
>>>
>>> So these things are done, but it is also possible that the specific
>>> quadtree functionality could be tested, in isolation from the
>>> projection. I think this may be oneo f the things Anthony is talking
>>> about -- answer testing can handle the big, complex items, and by
>>> breaking down to the fundamentals we can address isolated items from a
>>> unit testing perspective.
>>>
>>> >
>>> > So unfortunately, I think it would still come down to having a fake
>>> > frontend. It's not ideal, but it seems like any more isolation would
>>> > require
>>> > big rewrites to yt.
>>>
>>> One fun thing that is not usually known is that we have a fake
>>> frontend already, it just doesn't get used much. It's called the
>>> "Stream" frontend and it was designed originally to be used in
>>> ParaView, but now gets used by the (new, not-yet-documented/released)
>>> load_uniform_grid function as well as by Hyperion, the RT code by Tom
>>> R. It can set up AMR as well as static mesh. It's not terribly well
>>> documented, but there are examples on the wiki.
>>>
>>> One thing I've been thinking about is actually creating a couple fake
>>> outputs, which could be defined analytically with spheres of
>>> overdensity inside them. In principle, if we added refinement
>>> criteria, we could make this relatively complex data that was defined
>>> with only a few lines of code, but spun up a big in-memory dataset.
>>>
>>> (This exact thing is on my list of things to do and then to output in
>>> GDF, by the way...)
>>>
>>> That I think could come, down the road a bit. The refinement criteria
>>> wouldn't be too bad to implement, especially since we already have the
>>> grid splitting routines. I just don't think we should focus on it at
>>> the moment. But the uniform grid creation and loading works already
>>> -- I used it this morning. You can do it with:
>>>
>>> from yt.frontends.stream.api import load_uniform_grid
>>> ug = load_uniform_grid({"VelocityNorm":data1, "Density":data2}, [359,
>>> 359, 359], 1.0)
>>>
>>> the list is the dimensions of the data and the value is the to-cm
>>> conversion.
>>>
>>> >
>>> > Of course, I could be missing something. Matt, can you think of a better
>>> > way?
>>>
>>> I think for this specific example (and your damningly complex tracing
>>> of things through the source ...) the easiest thing to do is isolate
>>> the Cython routine, which it seems I was able to do only because I
>>> wrote it and which seems quite buried in the code, and to also provide
>>> high-level machinery for faking a frontend.
>>>
>>> -Matt
>>>
>>> >
>>> > - Casey
>>> >
>>> >
>>> > On Mon, Sep 24, 2012 at 11:02 AM, Anthony Scopatz <scopatz at gmail.com>
>>> > wrote:
>>> >>
>>> >> Helo Casey,
>>> >>
>>> >> Sorry for taking the whole weekend to respond.
>>> >>
>>> >>>> I would like to help with this, but it's difficult to figure out
>>> >>>> where
>>> >>>> to start.
>>> >>
>>> >>
>>> >> Not to worry. I think that any of the items listed at the bottom of
>>> >> Matt's
>>> >> original email
>>> >> would be a great place to start.
>>> >>
>>> >>>>
>>> >>>>
>>> >>>> Say I want to test projections. I make a fake 3D density field, maybe
>>> >>>> something as simple as np.arange(4**3).reshape((4, 4, 4)). I write
>>> >>>> down the
>>> >>>> answer to the x-projection. Now all I need to do is call
>>> >>>> assert_allclose(yt_result, answer, rtol=1e-15), but I don't know what
>>> >>>> pieces
>>> >>>> of low-level yt stuff to call to get to `yt_result`. Hopefully that's
>>> >>>> clear...
>>> >>>>
>>> >>>> Maybe this comes down to creating a fake frontend we can attach
>>> >>>> fields
>>> >>>> to?
>>> >>
>>> >>
>>> >> Actually, I disagree with this strategy, as I told Matt when we spoke
>>> >> last
>>> >> week.
>>> >> What is important is that we test the science and math parts of the
>>> >> code
>>> >> before, if ever, dealing with the software architecture that surrounds
>>> >> them.
>>> >>
>>> >> Let's taking your example of projections. What we need to test is the
>>> >> actual function
>>> >> or method which actually slogs through the projection calculation. In
>>> >> many cases in
>>> >> yt these functions are not directly attached to the front end but live
>>> >> in
>>> >> analysis, visualization
>>> >> or utilities subpackages. It is these such packages that we should
>>> >> worry
>>> >> about testing.
>>> >> We can easily create routines to feed them sample data.
>>> >>
>>> >> On the other hand, testing or mocking things like frontends should be a
>>> >> very low priority.
>>> >> At the end of the day what you are testing here is pulling in data from
>>> >> disk or other
>>> >> sources. Effectively, this is just re-testing functionality present in
>>> >> h5py, etc. That is not
>>> >> really our job. Yes, in a perfect world, front ends would be tested
>>> >> too.
>>> >> But I think that the
>>> >> priority should be placed on things like the KDTree.
>>> >>
>>> >> Be Well
>>> >> Anthony
>>> >>
>>> >>>>
>>> >>>>
>>> >>>> - Casey
>>> >>>>
>>> >>>>
>>> >>>> On Fri, Sep 21, 2012 at 2:42 PM, Matthew Turk <matthewturk at gmail.com>
>>> >>>> wrote:
>>> >>>>>
>>> >>>>> Hi all,
>>> >>>>>
>>> >>>>> As some of you have seen (at least Stephen), I filed a ticket this
>>> >>>>> morning about increasing testing coverage. The other night Anthony
>>> >>>>> and I met up in NYC and he had something of an "intervention" about
>>> >>>>> the sufficiency of answer testing for yt; it didn't take too much
>>> >>>>> work
>>> >>>>> on his part to convince me that we should be testing not just
>>> >>>>> against
>>> >>>>> a gold standard, but also performing unit tests. In the past I had
>>> >>>>> eschewed unit testing simply because the task of mocking data was
>>> >>>>> quite tricky, and by adding tests that use smaller bits we could
>>> >>>>> cover
>>> >>>>> unit testable areas with answer testing.
>>> >>>>>
>>> >>>>> But, this isn't really a good strategy. Let's move to having both.
>>> >>>>> The testing infrastructure he recommends is the nearly-omnipresent
>>> >>>>> nose:
>>> >>>>>
>>> >>>>> http://nose.readthedocs.org/en/latest/
>>> >>>>>
>>> >>>>> The ticket to track this is here:
>>> >>>>>
>>> >>>>>
>>> >>>>>
>>> >>>>> https://bitbucket.org/yt_analysis/yt/issue/426/increase-unit-test-coverage
>>> >>>>>
>>> >>>>> There are a couple sub-items here:
>>> >>>>>
>>> >>>>> 1) NumPy's nose test plugins provide a lot of necessary
>>> >>>>> functionality
>>> >>>>> that we have reimplemented in the answer testing utilities. I'd
>>> >>>>> like
>>> >>>>> to start using the numpy plugins, which include things like
>>> >>>>> conditional test execution, array comparisons, "slow" tests, etc
>>> >>>>> etc.
>>> >>>>> 2) We can evaluate, using conditional test execution, moving to nose
>>> >>>>> for answer testing. But that's not on the agenda now.
>>> >>>>> 3) Writing tests for nose is super easy, and running them is too.
>>> >>>>> Just
>>> >>>>> do:
>>> >>>>>
>>> >>>>> nosetest -w yt/
>>> >>>>>
>>> >>>>> when in your source directory.
>>> >>>>>
>>> >>>>> 4) I've written a simple sample here:
>>> >>>>>
>>> >>>>>
>>> >>>>>
>>> >>>>> https://bitbucket.org/yt_analysis/yt-3.0/src/da10ffc17f6d/yt/utilities/tests/test_interpolators.py
>>> >>>>>
>>> >>>>> 5) I'll handle writing up some mock data that doesn't require
>>> >>>>> shipping
>>> >>>>> lots of binary files, which can then be used for checking things
>>> >>>>> that
>>> >>>>> absolutely require hierarchies.
>>> >>>>>
>>> >>>>> --
>>> >>>>>
>>> >>>>> The way to organize tests is easy. Inside each directory with
>>> >>>>> testable items create a new directory called "tests", and in here
>>> >>>>> toss
>>> >>>>> some scripts. You can stick a bunch of functions in those scripts.
>>> >>>>>
>>> >>>>> Anyway, I'm going to start writing more of these (in the main yt
>>> >>>>> repo,
>>> >>>>> and this change will be grafted there as well) and I'll write back
>>> >>>>> once the data mocking is ready. I'd like it if we started
>>> >>>>> encouraging
>>> >>>>> or even mandating simple tests (and/or answer tests) for
>>> >>>>> functionality
>>> >>>>> that gets added, but that's a discussion that should be held
>>> >>>>> separately.
>>> >>>>>
>>> >>>>> The items on the ticket:
>>> >>>>>
>>> >>>>> * kD-tree for nearest neighbor
>>> >>>>> * Geometric selection routines
>>> >>>>> * Profiles
>>> >>>>> * Projections -- underlying quadtree
>>> >>>>> * Data object selection of data containers
>>> >>>>> * Data object selection of points
>>> >>>>> * Orientation class
>>> >>>>> * Pixelization
>>> >>>>> * Color maps
>>> >>>>> * PNG writing
>>> >>>>>
>>> >>>>> Is anyone willing to claim any additional items that they will help
>>> >>>>> write unit tests for?
>>> >>>>>
>>> >>>>> -Matt
>>> >>>>> _______________________________________________
>>> >>>>> yt-dev mailing list
>>> >>>>> yt-dev at lists.spacepope.org
>>> >>>>> http://lists.spacepope.org/listinfo.cgi/yt-dev-spacepope.org
>>> >>>>
>>> >>>>
>>> >>>>
>>> >>>> _______________________________________________
>>> >>>> yt-dev mailing list
>>> >>>> yt-dev at lists.spacepope.org
>>> >>>> http://lists.spacepope.org/listinfo.cgi/yt-dev-spacepope.org
>>> >>>>
>>> >>
>>> >>
>>> >> _______________________________________________
>>> >> yt-dev mailing list
>>> >> yt-dev at lists.spacepope.org
>>> >> http://lists.spacepope.org/listinfo.cgi/yt-dev-spacepope.org
>>> >>
>>> >
>>> >
>>> > _______________________________________________
>>> > yt-dev mailing list
>>> > yt-dev at lists.spacepope.org
>>> > http://lists.spacepope.org/listinfo.cgi/yt-dev-spacepope.org
>>> >
>>> _______________________________________________
>>> yt-dev mailing list
>>> yt-dev at lists.spacepope.org
>>> http://lists.spacepope.org/listinfo.cgi/yt-dev-spacepope.org
>>
>>
>>
>> _______________________________________________
>> yt-dev mailing list
>> yt-dev at lists.spacepope.org
>> http://lists.spacepope.org/listinfo.cgi/yt-dev-spacepope.org
>>
More information about the yt-dev
mailing list