Hello Peter, On Nov 28, 2013, at 4:56 AM, Peter Cock <p.j.a.cock@googlemail.com> wrote:
On Wed, Nov 27, 2013 at 9:02 PM, Greg Von Kuster <greg@bx.psu.edu> wrote:
Hello Peter,
I've done a significant amount of re-engineering of the Tool Shed's install and test framework over the past 2 weeks, so the results you are seeing are probably a reflection of some broken peices during the time frame in which I was making changes. I believe I have most of the things ironed out in 11545:32205f911e74 which is now running on the test tool shed.
We will be performing some test runs of thie new framework over the long weekend, so more correct test results should hopefully be showing up for yours and other repositories.
Hi Greg,
Things look much better today on the Test Tool Shed, with most of my tools under "Latest revision: all tool tests pass" again - or as expected a few under "Latest revision: missing tool tests" due to not yet having full test coverage.
I've only spotted one glitch so far, missing test results for: http://testtoolshed.g2.bx.psu.edu/view/peterjc/ncbi_blast_plus/b2795652d2b4
The above repository was not displaying the missing test components due to a problem in the previous test runs which I fixed yesterday. The problem resulted in test result data that could not be displayed. When I introduced the new feature for stroing and displaying multiple test runs, I decided to not eliminate the bad data, but just to let enough test runs proceed that the bad data would eventually be eliminated. I forced new test runs, and now things should be displaying as expected. Here is the display for the above repository for the latest run this morning: Automated tool test results Automated test environment Time tested: 2013-11-28 11:29:38 System: Architecture: Python version: Galaxy revision: Galaxy database version: Tool shed revision: 11549:5aede225ebef Tool shed database version: 21 Tool shed mercurial version: 2.2.3 Tools missing tests or test data Tool id: ncbi_rpstblastn_wrapper Tool version: 0.0.22 Tool guid: testtoolshed.g2.bx.psu.edu/repos/peterjc/ncbi_blast_plus/ncbi_rpstblastn_wrapper/0.0.22 Missing components: Functional test definitions missing for ncbi_rpstblastn_wrapper. Tool id: ncbi_blastdbcmd_info Tool version: 0.0.22 Tool guid: testtoolshed.g2.bx.psu.edu/repos/peterjc/ncbi_blast_plus/ncbi_blastdbcmd_info/0.0.22 Missing components: Functional test definitions missing for ncbi_blastdbcmd_info. Tool id: ncbi_makeblastdb Tool version: 0.0.22 Tool guid: testtoolshed.g2.bx.psu.edu/repos/peterjc/ncbi_blast_plus/ncbi_makeblastdb/0.0.22 Missing components: One or more test files are missing for tool ncbi_makeblastdb: four_human_proteins.fasta.log Tool id: ncbi_blastdbcmd_wrapper Tool version: 0.0.22 Tool guid: testtoolshed.g2.bx.psu.edu/repos/peterjc/ncbi_blast_plus/ncbi_blastdbcmd_wrapper/0.0.22 Missing components: Functional test definitions missing for ncbi_blastdbcmd_wrapper. Tool id: ncbi_rpsblast_wrapper Tool version: 0.0.22 Tool guid: testtoolshed.g2.bx.psu.edu/repos/peterjc/ncbi_blast_plus/ncbi_rpsblast_wrapper/0.0.22 Missing components: Functional test definitions missing for ncbi_rpsblast_wrapper.
I saw a recent commit on galaxy-central addressing a test time display - that may have explained why yesterday the Test Tool Shed was claiming the test results were from four days ago.
Yes, there is a flag that can be set for testing a repository that eliminates it from being tested by the install and test framework. This flag had been inappropriately set on many repositories, including the one to which you are referring. That problem has been corrected, so testing should be more stable.
I've also enhanced the "Tool test results" components to store a configurable number of test runs instead of just 1. The test tool shed is configured to store 5 test runs. So from here on you'll begin to see multiple containers in the Tool test results container - something like this:
That sounds useful when the framework itself is changing rapidly, or if there is anything stochastic in a test failure.
Presumably this shows the last 5 tests against the current tool revision?
It displays the last 5 test runs against any installable revision that is tested. We are currently restricting testing to the latest installable revision, so the 5 test runs will be displayed for that revision only.
Thank Greg, and enjoy Thanksgiving,
Thanks Peter, your too!
Peter