This is just a fork of coveralls with updated dependencies and replace request which is deprecated with form-data which was the library request was using for form posts. I also replace xo with eslint and prettier with google settings as it also used a bunch of deprecated dependencies.
Coveralls.io support for Node.js. Get the great coverage reporting of coveralls.io and add a cool coverage button (like the one above) to your README.
Add the latest version of coveralls
to your package.json:
npm install coveralls-next --save-dev
If you're using mocha, add mocha-lcov-reporter
to your package.json:
npm install mocha-lcov-reporter --save-dev
This script bin/coveralls.js
can take standard input from any tool that emits the lcov data format (including mocha's LCOV reporter) and send it to coveralls.io to report your code coverage there.
Once your app is instrumented for coverage, and building, you need to pipe the lcov output to ./node_modules/coveralls/bin/coveralls.js
.
This library currently supports Travis CI with no extra effort beyond piping the lcov output to coveralls. However, if you're using a different build system, there are a few necessary environment variables:
COVERALLS_SERVICE_NAME
(the name of your build system)COVERALLS_REPO_TOKEN
(the secret repo token from coveralls.io)COVERALLS_GIT_BRANCH
(the branch name)
There are optional environment variables for other build systems as well:
COVERALLS_FLAG_NAME
(a flag name to differentiate jobs, e.g. Unit, Functional, Integration)COVERALLS_SERVICE_NUMBER
(a number that uniquely identifies the build)COVERALLS_SERVICE_JOB_ID
(an ID that uniquely identifies the build's job)COVERALLS_SERVICE_JOB_NUMBER
(a number that uniquely identifies the build's job)COVERALLS_RUN_AT
(a date string for the time that the job ran. RFC 3339 dates work. This defaults to your build system's date/time if you don't set it)COVERALLS_PARALLEL
(set totrue
when running jobs in parallel, requires a completion webhook. More info here: https://docs.coveralls.io/parallel-build-webhook)
If you use this then there is no reason to have coveralls or coveralls-next library in your package as it has it's own npm version in the step. This doesn't use this library but the original coveralls npm package which will work just the same.
If you are using GitHub Actions CI, you should look into coverallsapp/github-action.
Parallel runs example workflow.yml
If you use this then there is no reason to have coveralls or coveralls-next library in your package as it has it's own npm version in the step. This doesn't use this library but the original coveralls npm package which will work just the same.
Here's our Orb for quick integration: coveralls/coveralls
Workflow example: config.yml
Parallel jobs example: .travis.yml
-
Install jest
-
Use the following to run tests and push files to coveralls on success:
jest --coverage && coveralls < coverage/lcov.info
Check out an example here which makes use of Travis CI build stages
-
Install blanket.js
-
Configure blanket according to docs.
-
Run your tests with a command like this:
NODE_ENV=test YOURPACKAGE_COVERAGE=1 ./node_modules/.bin/mocha \ --require blanket \ --reporter mocha-lcov-reporter | ./node_modules/coveralls/bin/coveralls.js
Instrumenting your app for coverage is probably harder than it needs to be (read here), but that's also a necessary step.
In mocha, if you've got your code instrumented for coverage, the command for a Travis CI build would look something like this:
YOURPACKAGE_COVERAGE=1 ./node_modules/.bin/mocha test -R mocha-lcov-reporter | ./node_modules/coveralls/bin/coveralls.js
Check out an example Makefile from one of my projects for an example, especially the test-coveralls build target. Note: Travis CI runs npm test
, so whatever target you create in your Makefile must be the target that npm test
runs (This is set in package.json's scripts
property).
istanbul cover ./node_modules/mocha/bin/_mocha --report lcovonly -- -R spec && cat ./coverage/lcov.info | ./node_modules/coveralls/bin/coveralls.js && rm -rf ./coverage
istanbul cover jasmine-node --captureExceptions spec/ && cat ./coverage/lcov.info | ./node_modules/coveralls/bin/coveralls.js && rm -rf ./coverage
Depend on nodeunit, jscoverage, and coveralls:
npm install nodeunit jscoverage coveralls-next --save-dev
Add a coveralls script to "scripts" in your package.json
:
"scripts": {
"test": "nodeunit test",
"coveralls": "jscoverage lib && YOURPACKAGE_COVERAGE=1 nodeunit --reporter=lcov test | coveralls"
}
Ensure your app requires instrumented code when process.env.YOURPACKAGE_COVERAGE
variable is defined.
Run your tests with a command like this:
npm run coveralls
For detailed instructions on requiring instrumented code, running on Travis CI and submitting to coveralls see this guide.
Client-side JS code coverage using PhantomJS, Mocha and Blanket:
-
Configure Mocha for browser
-
Mark target script(s) with
data-cover
HTML attribute -
Run your tests with a command like this:
./node_modules/.bin/poncho -R lcov test/test.html | ./node_modules/coveralls/bin/coveralls.js
lab -r lcov | ./node_modules/.bin/coveralls
Works with almost any testing framework. Simply execute
npm test
with the nyc
bin followed by running its reporter:
nyc npm test && nyc report --reporter=text-lcov | coveralls
Simply run your tap tests with the COVERALLS_REPO_TOKEN
environment
variable set and tap will automatically use nyc
to report
coverage to coveralls.
Usage: coveralls.js [-v] filepath
-v
,--verbose
filepath
- optionally defines the base filepath of your source files.
If you're running locally, you must have a .coveralls.yml
file, as documented in their documentation, with your repo_token
in it; or, you must provide a COVERALLS_REPO_TOKEN
environment variable on the command-line.
If you want to send commit data to coveralls, you can set the COVERALLS_GIT_COMMIT
environment-variable to the commit hash you wish to reference. If you don't want to use a hash, you can set it to HEAD
to supply coveralls with the latest commit data. This requires git to be installed and executable on the current PATH.
I generally don't accept pull requests that are untested or break the build, because I'd like to keep the quality high (this is a coverage tool after all!).
I also don't care for "soft-versioning" or "optimistic versioning" (dependencies that have ^, x, > in them, or anything other than numbers and dots). There have been too many problems with bad semantic versioning in dependencies, and I'd rather have a solid library than a bleeding-edge one.