I'm working on a small gem and included simplecov to spec_helper.rb two lines:
require 'simplecov'
SimpleCov.start
When I run the rspec test, the simplecov seems started correctly but the report is not:
Finished in 0.00214 seconds
8 examples, 0 failures
Coverage report generated for /home/......
spec to /home/megas/Work/calc/coverage. 0 / 0 LOC (0.0%) covered.
What might be a problem and how to fix it? Thanks
Also make sure to enable simplecov (a.k.a. SimpleCov.start) at the very beginning of your file; especially before you require your code.
I had the same symptoms. My problem was that in my test file:
#spec/oneclass_spec.rb
require 'oneclass'
require 'spec_helper'
...Rest of the test file
And I needed to change the order of the requires to:
#spec/oneclass_spec.rb
require 'spec_helper'
require 'oneclass'
...Rest of the test file
Hope this helps someone, I was going crazy...
I had a similar issue. For some reason, some modules were reported with 0% of coverage. After some investigation, I found that one of the initialisers required a controller, which caused modules to be loaded before Simplecov. What I did I moved Simplecov to initialiser:
# config/initializers/_coverage_rspec.rb
if Rails.env.test?
require 'simplecov'
SimpleCov.start 'rails'
end
If you're using spring, remember to turn it off when you run tests with coverage. In another case, the complete application code will be loaded before SimpleCov and report 0% coverage.
Just in case the above two answers didn't work (as in my case) a user on simplecov's github issues page suggested this, which worked for me.
Add this after you require simplecov-
module SimpleCov::Configuration
def clean_filters
#filters = []
end
end
SimpleCov.configure do
clean_filters
load_adapter 'test_frameworks'
end
If one of the above didn't work.
verify that in test.rb:
config.eager_load = false
In my case the issue was spring - I had to create a config/spring.rb with the following:
if ENV['RAILS_ENV'] == 'test'
require 'simplecov'
SimpleCov.start
end
as documented here.
I am running scripts from the command line and I found the solution was simply to put an exit at the end of my script. Doh!
Alternatively, the following also works
SimpleCov.at_exit do
SimpleCov.result.format!
end
Related
I am trying to add SimpleCov coverage to my gem that has a binary.
I would like to test its command line interface, so I wish to test the binary execution itself, and not the library it uses.
I am getting no coverage report (0 LOC) by SimpleCov.
As I understand, the issue is most likely due to the fact that in my test (either cucumber features or rspec specs), I am executing the gem's binary with system or popen3, but I don't know how I can tell SimpleCov to "follow through" (or if I am barking at the right tree...).
I have tried playing with SimpleCov.command_name, SimpleCov.pid = $$, SimpleCov.track_files and almost every other remotely related configuration I found.
I do not wish to use Aruba, although I have tried reviewing their source to look for possible answers.
Related code snippets:
# spec_helper.rb
require 'simplecov'
SimpleCov.start
require 'rubygems'
require 'bundler'
Bundler.require :default, :development
# test_spec.rb
require 'spec_helper'
describe "my bin" do
it "should be covered" do
system 'bin/runme'
end
end
I have prepared a minimal repo as an easy testing ground, if that helps at all.
well it's not easy to find a solution to this issue in particular.
not sure this solves your problem but maybe a start?
without the start block which modifies the filters i could not get SimpleCov to watch the bin directory (using the sample github repo you provided).
Used command_name to give the main processes coverage reports a name and then in the fork used command_name to give the forked processes report a name (SimpleCov merges them for us as long as they have different names).
then used load to load the bin file instead of using system.
(I couldn't figure out a way to make system or spawn add to the coverage reports, maybe if you called it through a script that restarts SimpleCov for you with an alternate command_name)
again, not sure if this is exactly what you are looking for but may be a start. code below:
# spec_helper.rb
require 'simplecov'
SimpleCov.command_name "main_report"
SimpleCov.start do
filters.clear # This will remove the :root_filter and :bundler_filter that come via simplecov's defaults
add_filter do |src|
!(src.filename =~ /^#{SimpleCov.root}/) unless src.filename =~ /bin/ #make sure the bin directory is allowed
end
end
require 'rubygems'
require 'bundler'
Bundler.require :default, :development
# test_spec.rb
require 'spec_helper'
describe "my bin" do
it "should be covered" do
pid = Process.fork do
SimpleCov.start do
command_name "bin_report_section"
end
load "bin/runme"
end
end
end
result:
Coverage report generated for bin_report_section, main_report to
/home/korreyd/simplecov-debug/coverage.
1 / 1 LOC (100.0%) covered.
Have you tried this? https://blog.yossarian.net/2018/04/01/Code-coverage-with-Simplecov-across-multiple-processes
Basically, in your spec_helper.rb
if ENV["COVERAGE"]
require "simplecov"
# Only necessary if your tests *might* take longer than the default merge
# timeout, which is 10 minutes (600s).
SimpleCov.merge_timeout(1200)
# Store our original (pre-fork) pid, so that we only call `format!`
# in our exit handler if we're in the original parent.
pid = Process.pid
SimpleCov.at_exit do
SimpleCov.result.format! if Process.pid == pid
end
# Start SimpleCov as usual.
SimpleCov.start
end
Then inside your bin/runme, add:
if ENV["COVERAGE"]
# Give our new forked process a unique command name, to prevent problems
# when merging coverage results.
SimpleCov.command_name SecureRandom.uuid
SimpleCov.start
end
Child process' test coverage will merge into the parent's process.
If you use SimpleCov's coverage_dir, make sure it's in all SimpleCov.start blocks so that results are written to same location.
I am creating a test automation tool that runs a rspec test from within ruby, not from command line.
I run my test like this that runs and gives me a output hash with pass / fail etc.
config = RSpec.configuration
json_formatter = RSpec::Core::Formatters::JsonFormatter.new(config.out)
reporter = RSpec::Core::Reporter.new(json_formatter)
config.instance_variable_set(:#reporter, reporter)
RSpec::Core::Runner.run(["#{Rails.root}/spec/test1_spec.rb"])
puts json_formatter.output_hash
All is great until i run it again and i get the error
> undefined method `visit' for #<RSpec::Core::ExampleGroup::Nested_7:0x007ff3d343d978>
how do i reload / make sure all spec_helper and files are loaded before i run it again ?
thanks
Rick
Add
require 'spec_helper'
to the top of each spec file.
I had the same problems, turns out it can be fixed with one line of code only:
require 'capybara'
include Capybara::DSL # Add this line right after requiring capybara
Source: http://codedecoder.wordpress.com/2013/01/10/undefined-method-visit-for-rspec-capybara/
I had a similar issue and had to do a RSpec#reset between runs to make sure everything was cleaned up.
http://www.rubydoc.info/github/rspec/rspec-core/RSpec#reset-class_method
I have a requirement where in, i need to get the coverage so far.
If I stop the server, the report gets generated and I do get the coverage so far. But if i start the server again, my previous coverage results are lost and I can only get the coverage after the server was restarted.
Is there a way for me to keep checking periodically for the the coverage% - without stopping the server?
If i try to generate a report without starting the server, by using the following command, in rails console (SimpleCov.result.format! ),I dont get anycoverage number.
The following is my config in my config/boot.rb:
require 'simplecov'
SimpleCov.start 'rails' do
add_filter "/vendor/"
end
Please share your thoughts
Thanks
Ramya
This is the content of my boot.rb:
require 'simplecov'
# # create coverage directory if it doesn't exist already.
Dir.mkdir("coverage") unless Dir.exist?("coverage")
SimpleCov.start 'rails' do
SimpleCov.use_merging(true)
add_filter "/vendor/"
SimpleCov.merge_timeout 30
end
require 'rubygems'
# Set up gems listed in the Gemfile.
ENV['BUNDLE_GEMFILE'] ||= File.expand_path('../../Gemfile', __FILE__)
require 'bundler/setup' if File.exists?(ENV['BUNDLE_GEMFILE'])
The pre-requisites for SimpleCov to work properly are documented here: Getting Started with SimpleCov. You must be having the SimpleCov related code inside the boot.rb file after the Rails loading code. This is wrong. Promote all that code to the top and the SimpleCov.result.format! method will work inside the console.
However, it's generally a bad idea to have any extra code inside the boot.rb. Usually, coverage reports are needed only in the test environment (when the code is committed and a continuous integration server like Travis runs the full test suite and generates a coverage report). Hence, the documentation refers to this style of setup where everything related to SimpleCov runs in the test environment. The first topic in the Getting Started section mentions that you need to have the SimpleCov.start line at the beginning of the test_helper file (spec_helper.rb if you're using Rspec) since that is the file that loads the Rails environment; which means that you end up loading SimpleCov and it's configuration before loading the actual application code and you get a correct output.
require 'simplecov'
SimpleCov.start do
coverage_dir "custom-coverage_"+Time.now.strftime("%m_%d_%Y").to_s
end
rubocop is a code style checker for Ruby. A similar tool to rubocop, Cane, can be integrated with Rake. I prefer rubocop to Cane since rubocop makes checks based on the Ruby Style Guide and it seems to spot more problems. To automate the process of style checking I would like to integrate rubocop with Rake so that the build fails if code quality is lacking.
Gem already supports adding tests to packages via Rake. I would like to do the same with style checks so that style checks are run along with the tests. How can I do this?
If it helps to start with a Rakefile here is one:
# -*- coding: utf-8; mode: ruby -*-
require 'bundler/gem_tasks'
require 'rake/testtask'
Rake::TestTask.new do |t|
t.libs << 'test'
t.test_files = FileList['test/unit/test*.rb']
end
desc 'Run tests'
task default: :test
As of version 0.10.0 rubocop contain a custom rake task that you can use. Just put the following in your Rakefile
require 'rubocop/rake_task'
RuboCop::RakeTask.new
Make sure to use upper-case 'R' and 'C' or you will get a NameError.
I highly recommend,
require 'rubocop/rake_task'
RuboCop::RakeTask.new(:rubocop) do |t|
t.options = ['--display-cop-names']
end
This uses the rubocop's own rake tasks and allows you to pass options if you like.
You will probably find https://github.com/yujinakayama/guard-rubocop useful if you use Guard for your RSpec tests. It enables Rubocop to give you instant feedback as soon as you save the file, along with your test results.
I needed to do something similar myself, and ended up looking in the internal source code of the RuboCop::RakeTask here:
https://github.com/rubocop/rubocop/blob/a34a1c2c2dd1fa6d90ffd06c183421a495a0717c/lib/rubocop/rake_task.rb#L40-L43
require 'rubocop'
cli = CLI.new
puts 'Running RuboCop...' if verbose
result = cli.run(options)
abort('RuboCop failed!') if result.nonzero? && fail_on_error
You can actually invoke similar code directly in your own codebase / rake task.
I ended up writing a little wrapper module I can call to, with some default flags that I always want to be applied:
module RubocopCli
def self.run!(*args)
require "rubocop"
cli = RuboCop::CLI.new
result = cli.run(["--display-cop-names", "--force-exclusion", "--fail-level", "autocorrect", *args])
raise "RubocopCli.run! Linting failed." if result.nonzero?
end
end
Then you can call it with additional args from any task, or app code, like:
files_to_lint = %w[lib/whatever.rb spec/lib/whatever_spec.rb]
RubocopCli.run!("--auto-correct", *files_to_lint)
You can shell out via Rake with the options you prefer:
desc 'Run Rubocop with options'
task rubocop: :environment do
sh 'bundle exec rubocop -D --format offenses --format progress || true'
end
I then recommend modifying the default task to include the output. The trick is to clear the task and then add back what you want. Note the need to end with || true so that an error from Rubocop will not prevent the next task from running. Here's what I do, which also uses parallel tests:
task(:default).clear.enhance ['parallel:parallel_prepare', 'parallel:spec',
'parallel:features', 'lint:rubocop',
'lint:rails_best_practices']
I would recommend shelling out to the rubocop program. It's the simplest solution. Just add this to your Rakefile:
task test: :rubocop
task :rubocop do
sh 'rubocop'
end
I have been following the 15 TDD steps to create a Rails application guide - but have run into an issue I cannot seem to resolve. For the functional test of the WordsController, I have the following code:
class WordsControllerTest < ActionController::TestCase
test "should get learn" do
get 'learn'
assert_response :success
end
test "learn passes a random word" do
some_word = Word.new
Word.expects(:random).returns(some_word)
get 'learn'
assert_equal some_word, assigns('word')
end
end
In the Word class I have the following code:
class Word < ActiveRecord::Base
def self.random
all = Word.find :all
all[rand(all.size)]
end
end
When I run the tests, I experience the following error (shortened for brevity):
1) Failure: unexpected invocation: Word(...).random() satisfied expectations:
- expected exactly once, already invoked once: Word(...).random()
I have tried changing changing the order of the tests along with a multitude of other things, but time and time again I continue to receive the same test failure - that Word.random() has already been invoked.
I'm running Rails 3.0 beta 4 and Mocha 0.9.8. I've searched long and hard for a solution to my problem, but I can't seem to find it. I'm new to Ruby/Rails so am rather unfamiliar with the language and the frameworks.
Thanks in advance!
mocha needs to be loaded last. I struggled a lot with this problem too.
#Gemfile
group :test
gem 'mocha', '~>0.9.8', :require => false
...
end
and
test_helper.rb
....
#at the very bottom
require 'mocha'
I had the same problem, mocked functionality was not isolated to a test, it seems to be a problem with the load order of Mocha.
I had some issues getting Mocha to work with Rails3. I found a few stackoverflow posts regarding, but didn't stumble across the solution until I found a post on agoragames.com
Basically, in the Gemfile of your project, the require for Mocha should look like:
gem 'mocha', :require => false
Then in test/test_helper.rb, add a require line for mocha:
...
...
require File.expand_path('../../config/environment', __FILE__)
require 'rails/test_help'
require 'mocha'
class ActiveSupport::TestCase
...
...
I think the require line for mocha in the Gemfile means that you need to already have mocha installed as a gem, bundler won't take care of it for you.
How are you requiring mocha? Are you using bundler? It sounds a bit as if the mocha teardown hook isn't being called?
Additionally, it seems mocha_teardown is not being called with rails31. Mocks that are setup are never removed... (this additional hack fixes it)
class ActiveSupport::TestCase
def teardown
super
Mocha::Mockery.instance.teardown
Mocha::Mockery.reset_instance
end
end
Those solutions didn't work for me on their own, using Ruby 2.2.2, Rails 4.2.2, mocha 1.1.0, shoulda-context 1.2.1, factory_girl_rails 4.5.0 and a few more testing related gems.
What did it was also moving these two lines at the bottom of my test_helper.rb:
require 'mocha/setup'
require 'mocha/test_unit'
I also removed require 'test/unit'. It appears that mocha/test_unit already does that for me.