The contest, recently launch in California for the first time and currently accepting entries, is an invitation for youth (age 18 and younger) to go outside (or to a zoo, aquarium, or museum), to "get to know" their wild neighbors, and then to share their experience with the world by creating an original work of art, photography, or writing and then entering their creation in the contest.
Thursday, October 15, 2009
Kids' contest: art and nature
http://www.gettoknow.ca/us/contest/
...Perhaps something EOL could get involved with? Sounds like it's CA-only at the moment, but I could be mistaken.
Thursday, July 16, 2009
Legacy Doesn't Always Mean "Old"
I've decided the EOL website development project is a "Legacy Code" project.
That may seem a bit naive, since it's really a start-up project that's two years old, give or take. However, there are three reasons for my decision.
I found these to be very effective! To answer the first one, we developed the following three answers:
That may seem a bit naive, since it's really a start-up project that's two years old, give or take. However, there are three reasons for my decision.
- Because it was a rapid start-up project that was done without "real" Agile development. It accrued a surprising amount of code debt in its early days.
- Because the database is non-Rails in some serious ways. Particularly in its use of compound keys and the fact that it's spread over three databases instead of one. There are good business reasons for these decisions, but they affect our use of Rails.
- Because of the lack of robust testing.
- What are our overall testing goals, really?
- How well are we achieving them, currently?
- What are the most pressing needs?
- How can we best apply our time to improve tests?
- Facilitate careful development (TDD).
- Prevent regressions from being deployed.
- Create enough trust to allow refactoring.
- View and controller tests are really much like "unit" tests, since they are testing a method on a class... it's just a different "kind" of class.
- It's okay to throw away unit tests (including controller/view), as long as the features they are supporting are all tested with higher-level tests, and those tests are passing. Implementation methods change, some of the requirements of the past are likely to change, too.
- Try to write tests that are not fragile (meaning: refactoring doesn't fail tests). The underlying idea is that when a feature is complete, all tests should pass and when the feature is incomplete, some tests should fail. Period. The interals are less important, though someone may have a need to test them (at the Unit level) to help with developement.
- The blackbox, factory approach is more appropriate than mocks and specs in EOL's environment. But we need to make the domain logic approachable to developers (new ones in particular). Copy/pasting solutions from other tests /works/, but is not ideal. We need to make this easier to use, more convenient. Developers should /want/ to write tests, becaise it's easy to do. Newcomers have varying levels of domain knowledge.
Friday, July 10, 2009
On Creating Effective Web Interfaces
At EOL, we've been asked to implement some new features that we (the developers) didn't like. They didn't seem to adhere to web-standards, didn't look particularly good, and didn't really seem to capture the problem (caveat: as we see it) elegantly.
This sparked a discussion (healthy, not heated) about how we could best approach design and web usability as we add features. Another developer wrote a blog-post about it. (...internal-only; sorry, no link for you!) I felt compelled to talk about the same subject myself and field some ideas. His post was all well-formatted, edited, and technical. Mine is casual and sloppy. So it goes. ; )
So what's the problem?
We're using Agile methods, so the proximal problem--which we are already trying to address--is that the temptation is strong to just "get something out there". Users can respond to it, and we can quickly adjust our approach to build something they will like better.
This has a few negative consequences:
The core problem, I think, is that we really need someone skilled at web interfaces. But we don't have the money to add someone to the team to do this, and everyone currently on the team is too busy to pick up the skill and/or apply it at the expense of their current duties.
In short, we are left with only one option: establish some guidelines by which we can do the best we can, on our own.
To that end, I like the idea that my associate mentioned (interestingly enough, by quoting yet another associate of mine): a list of usability questions that can be asked to gauge whether an interface works or not.
...I think this is a reasonable start. That said:
This sparked a discussion (healthy, not heated) about how we could best approach design and web usability as we add features. Another developer wrote a blog-post about it. (...internal-only; sorry, no link for you!) I felt compelled to talk about the same subject myself and field some ideas. His post was all well-formatted, edited, and technical. Mine is casual and sloppy. So it goes. ; )
So what's the problem?
We're using Agile methods, so the proximal problem--which we are already trying to address--is that the temptation is strong to just "get something out there". Users can respond to it, and we can quickly adjust our approach to build something they will like better.
This has a few negative consequences:
- The code gets ugly and risks leaving remnants of old styles when it changes. Have a look at our CSS, for example. :(
- I think users can get scared away. At least for me, I know I've tried to use some feature of, say, Yahoo! Mail, hated it, and never looked at it again. ...They may very well have cleaned things up later, but I don't care to look.
- User feedback tends to... well... suck. In my experience, they'll ask for things that actually result in clutter. Also, it's only the squeaky wheels that get heard, and their needs may be vastly different from the majority.
- There is a big but poorly-understood difference between a feature that's easy to use and one that is "intuitive". The latter implies that a newcomer can step in and use the tool... well... intuitively. This is great! But it often creates an interface that wears on your quickly, or ends up something you hate over time. Photoshop's clone tool is a good example of a tool that's easy to use, but not at all intuitive. It's important to know which is better for your problem. I think that user-generated feedback generally comes from users who have tinkered with an interface for something on the order of minutes. They're talking about intuitive use, not easy use.
The core problem, I think, is that we really need someone skilled at web interfaces. But we don't have the money to add someone to the team to do this, and everyone currently on the team is too busy to pick up the skill and/or apply it at the expense of their current duties.
In short, we are left with only one option: establish some guidelines by which we can do the best we can, on our own.
To that end, I like the idea that my associate mentioned (interestingly enough, by quoting yet another associate of mine): a list of usability questions that can be asked to gauge whether an interface works or not.
...I think this is a reasonable start. That said:
- I don't know who we should ask these questions.
- I worry that we'll end up with design-by-committee interfaces, which usually suck. Ideally, I hope that design can be limited to two people on any given feature. ...Ideally-ideally, the same two people on every feature, to maintain consistency.
- Nothing in the questions mentions web standards/best-practices.
- I worry that this is going to take up more time than we have, and breed frustration at the slow implementation of new features.
- I think the questions are good for measuring usability, but there are other concerns, such as intuitive-vs-easy, elegance, and the "message" conveyed by each feature. I hope we can add questions into the mix so that these things are considered.
- I think it's worth investing two or three people's time over a day or two (each) to research usability, and contribute to the "checklist" that we're building. This could include best-practices, design methodology, questions to ask, and the like.
Tuesday, June 2, 2009
"rake stats" should count all my stuff!
We have some custom directories for Specs, which aren't counted when you run "rake stats". I thought we deserved more credit! : )
I put at the bottom of our Rakefile. It's a little... brute-force, as it trumps RSpec's own task of the same name, but it does the trick:
I put at the bottom of our Rakefile. It's a little... brute-force, as it trumps RSpec's own task of the same name, but it does the trick:
# We have some pretty customized stat directories, so:
require 'spec/rake/spectask'
namespace :spec do
# Setup specs for stats
task :statsetup do
require 'code_statistics'
::STATS_DIRECTORIES << %w(Blackbox\ specs spec/blackbox) if File.exist?('spec/blackbox')
::STATS_DIRECTORIES << %w(Model\ specs spec/models) if File.exist?('spec/models')
::STATS_DIRECTORIES << %w(View\ specs spec/views) if File.exist?('spec/views')
::STATS_DIRECTORIES << %w(Controller\ specs spec/controllers) if File.exist?('spec/controllers')
::STATS_DIRECTORIES << %w(Selenium\ specs spec/selenium) if File.exist?('spec/selenium')
::STATS_DIRECTORIES << %w(Helper\ specs spec/helpers) if File.exist?('spec/helpers')
::STATS_DIRECTORIES << %w(Library\ specs spec/lib) if File.exist?('spec/lib')
::STATS_DIRECTORIES << %w(Cucumber\ features features) if File.exist?('features')
::CodeStatistics::TEST_TYPES << "Model specs" if File.exist?('spec/models')
::CodeStatistics::TEST_TYPES << "View specs" if File.exist?('spec/views')
::CodeStatistics::TEST_TYPES << "Blackbox specs" if File.exist?('spec/blackbox')
::CodeStatistics::TEST_TYPES << "Selenium specs" if File.exist?('spec/selenium')
::CodeStatistics::TEST_TYPES << "Controller specs" if File.exist?('spec/controllers')
::CodeStatistics::TEST_TYPES << "Helper specs" if File.exist?('spec/helpers')
::CodeStatistics::TEST_TYPES << "Library specs" if File.exist?('spec/lib')
::CodeStatistics::TEST_TYPES << "Cucumber features" if File.exist?('features')
::STATS_DIRECTORIES.delete_if {|a| a[0] =~ /test/}
end
end
Friday, May 15, 2009
RCov Gem Installation problem
I couldn't get RCov installed this week:
$ sudo gem install rcov --tracePoking around a wee bit, I discovered that this was just rubyforge being obnoxious. Go download the tarball form of the file and then just "sudo gem install [name of the tarball]". All good.
ERROR: While executing gem ... (OptionParser::InvalidOption)
invalid option: --trace
Friday, March 20, 2009
Testing Your WebApp
I've been looking into the various ways to test web applications, including performance, acceptance tests, and cross-browser compatibility. In my travels over the past week-and-a-half, I have decided there are three levels to this:
A good first step is to validate with the W3C tools and try to stay 100% standards compliant. But you're doing that already, I'm sure. : )
Adobe has a neat "Meer Meer" application that's coming... sometime. There is also SuperPreview. Both of these allow cross-browser design using nothing more than whatever web-browser you have installed: they call the real browsers on the server-side, take snapshots, and serve you the images. The coolest feature here is that you can overlay two browser's snapshots and have the differences jump out at you. ...Both of these are unreleased, but just around the corner. Very cool.
If snapshots are enough, you can get these cheap: $50 a month for Litmus. You can hook these tests into your testing suite. CrossBrowserTesting has a demo video here. Nothing earth-shattering: just note that you have to do these "tests" by hand.
IBM says: "For instance, for a monthly fee, Browsercam.com gives you access to a broader range of platforms and browsers—including some mobile platforms—and these pages are dynamic, allowing you to test Ajax apps and other JavaScript-based pages. Some services, like Litmusapp, can integrate with your testing and development workflows, offering versioning support and ways to publish test results to external clients."
Interestingly, if you want to do these things by hand, there are free Microsoft images of common Windows versions/Browser combos for use by developers. The bummer is that you have to renew them every so often (and only supported OS/browsers are represented), but it's a great way to test your app without spending too much money on setup.
A decent rundown of the "images of site" tools is here; (there were other overviews, not as good) we should see what the JavaScript-capable options are really capable of. If you need older versions, you'll have to install those yourself, but old IE versions are still available on the web (not from Microsoft, however).
Functional (Acceptance) Testing
For acceptance testing your Javascript, there are a few choices in the world of headless (ie: much faster) browsers. HtmlUnit seems a usable Java "mock" browser. There's a JRuby wrapper for it called Celerity, which could get me started faster. Similar is jsTest, also in Java. Jiffy struck me as overly complex, and hard-wired to Oracle for measures. Joe Dog Siege looks like a usable, mostly command-line driven load tester.
An interesting but probably less-useful alternative is to just check the JS for problems using crosscheck, but as of now it doesn't seem to have usable documentation, so...
If you don't mind (or have a need to be) using real browsers, Selenium is the standard right now for writing acceptance tests, and it's really easy to set up. Whether it adds overhead is an interesting question; I would assume yes based on some comments from other tool's sites. That said, A) You don't want to mix your acceptance tests with your performance tests, and B) it seems to be a standard right now, and I do recommend using it because of that. There are some alternatives, though. For example, JsUnit is a way to do the tests in native JavaScript, which is quite possibly useful.
Performance Testing
"Performance testing" (actually lots of sub-categories; read on) is a rather large arena, and I found this stuff quite interesting. ; ) I found the best introduction to the topic to be this neat video from Google, which gives some definitions:
Some other tips included: run every test at least 5 times before you measure it: inconsistency must be accounted for. She recommends a single five-minute test run "on demand", when a developer requests it, because you get *some* information from it, but nightlies are still required.
She recommends JMeter to measure loads (but it doesn't handle JavaScript) (There is a demo of how to get started, too), or Grinder (for "APIs").
Note that you should track not only the QPS, response times, etc, but also the CPU load, memory, versions of every piece of software on the machine, and disk i/o.
Interestingly, she says that Linux is typically performant as long as you don't exceed any single resource's 80% mark.
You can buy load testing. Prices on LoadStorm, for example, range from $30 to $200 a month. LoadRunner is another alternative (software that you use), by HP, which appears to be rather expensive... (Well, when they don't list a price, one must assume...)
BrowserStorm looks reasonable for a paid service, and has an explanation of its features in a blatant plug on Ajaxian.
Hope this helps.- Testing the design of your site. Does it look similar enough in all supported browsers?
- Testing the function of your site. Does all the Javascript work? Is everything that needs to render rendering?
- Testing the reliability of your site. How fast is it? How many users can it handle? For how long?
A good first step is to validate with the W3C tools and try to stay 100% standards compliant. But you're doing that already, I'm sure. : )
Adobe has a neat "Meer Meer" application that's coming... sometime. There is also SuperPreview. Both of these allow cross-browser design using nothing more than whatever web-browser you have installed: they call the real browsers on the server-side, take snapshots, and serve you the images. The coolest feature here is that you can overlay two browser's snapshots and have the differences jump out at you. ...Both of these are unreleased, but just around the corner. Very cool.
If snapshots are enough, you can get these cheap: $50 a month for Litmus. You can hook these tests into your testing suite. CrossBrowserTesting has a demo video here. Nothing earth-shattering: just note that you have to do these "tests" by hand.
IBM says: "For instance, for a monthly fee, Browsercam.com gives you access to a broader range of platforms and browsers—including some mobile platforms—and these pages are dynamic, allowing you to test Ajax apps and other JavaScript-based pages. Some services, like Litmusapp, can integrate with your testing and development workflows, offering versioning support and ways to publish test results to external clients."
Interestingly, if you want to do these things by hand, there are free Microsoft images of common Windows versions/Browser combos for use by developers. The bummer is that you have to renew them every so often (and only supported OS/browsers are represented), but it's a great way to test your app without spending too much money on setup.
A decent rundown of the "images of site" tools is here; (there were other overviews, not as good) we should see what the JavaScript-capable options are really capable of. If you need older versions, you'll have to install those yourself, but old IE versions are still available on the web (not from Microsoft, however).
Functional (Acceptance) Testing
For acceptance testing your Javascript, there are a few choices in the world of headless (ie: much faster) browsers. HtmlUnit seems a usable Java "mock" browser. There's a JRuby wrapper for it called Celerity, which could get me started faster. Similar is jsTest, also in Java. Jiffy struck me as overly complex, and hard-wired to Oracle for measures. Joe Dog Siege looks like a usable, mostly command-line driven load tester.
An interesting but probably less-useful alternative is to just check the JS for problems using crosscheck, but as of now it doesn't seem to have usable documentation, so...
If you don't mind (or have a need to be) using real browsers, Selenium is the standard right now for writing acceptance tests, and it's really easy to set up. Whether it adds overhead is an interesting question; I would assume yes based on some comments from other tool's sites. That said, A) You don't want to mix your acceptance tests with your performance tests, and B) it seems to be a standard right now, and I do recommend using it because of that. There are some alternatives, though. For example, JsUnit is a way to do the tests in native JavaScript, which is quite possibly useful.
Performance Testing
"Performance testing" (actually lots of sub-categories; read on) is a rather large arena, and I found this stuff quite interesting. ; ) I found the best introduction to the topic to be this neat video from Google, which gives some definitions:
- performance testing - how fast will it respond under load X? You need to test things at least long enough to reach a steady state. Start up slowly: don't just throw 100 users at it, build it up over a warm-up period. The machines under load periodically push performance dat (see blow) to monitoring servers, which analyze and store the data. These monitoring servers are also monitoring themselves (so you know what the effect of storing the data is). To the extent possible, remove the network: run everything in its own little environment.
- stress testing - at what point does the system break under load (and does it do so gracefully)?
- load (or liability) testing - given a high load (80% of max) for an extended period of time (Google does 1 month, 72 hours is a minimum recommendation), how does it perform? Often there are cron-related or system-level tasks (ex: log rotation) that affect performance only at certain times, so this needs to be examined.
- benchmark - a simplified measurement that is repeatable, representing a typical customer's activity on the system.
- profiling - When you change some aspect of the system (memory, load balancer, etc) and measure the differences in benchmark performance.
- availability testing - when a system fails, how long does it take to come back up? Google operates under the assumption that their equipment sucks, so they will solve that problem in software.
Some other tips included: run every test at least 5 times before you measure it: inconsistency must be accounted for. She recommends a single five-minute test run "on demand", when a developer requests it, because you get *some* information from it, but nightlies are still required.
She recommends JMeter to measure loads (but it doesn't handle JavaScript) (There is a demo of how to get started, too), or Grinder (for "APIs").
Note that you should track not only the QPS, response times, etc, but also the CPU load, memory, versions of every piece of software on the machine, and disk i/o.
Interestingly, she says that Linux is typically performant as long as you don't exceed any single resource's 80% mark.
You can buy load testing. Prices on LoadStorm, for example, range from $30 to $200 a month. LoadRunner is another alternative (software that you use), by HP, which appears to be rather expensive... (Well, when they don't list a price, one must assume...)
BrowserStorm looks reasonable for a paid service, and has an explanation of its features in a blatant plug on Ajaxian.
Monday, March 2, 2009
AIM Chatroom Death
...So we EOL developers hang out in an AOL chatroom. Last week, I got logged out of it, and couldn't get back in. ...Not sure why. I hoped the problem would just go away, but it never resolved itself. The error I was getting was:
Useless message! But eventually I found a page describing the problem, with the suggestion to change the AIM server port from 5190 to 443.
That worked.
Couldn't join chatroom "".
An AIM service error occurred.
Error: Serv:RequestTimeout
An AIM service error occurred.
Error: Serv:RequestTimeout
ChatError:NoError
Useless message! But eventually I found a page describing the problem, with the suggestion to change the AIM server port from 5190 to 443.
That worked.
Subscribe to:
Posts (Atom)