Node.js idiomatic client for Google Cloud Platform services.
This client supports the following Google Cloud Platform services at a General Availability (GA) quality level:
- Cloud Datastore (GA)
- Cloud Natural Language (GA)
- Cloud Storage (GA)
- Cloud Translation API (GA)
- Google Stackdriver Logging (GA)
This client supports the following Google Cloud Platform services at a Beta quality level:
- Cloud Data Loss Prevention (Beta)
- Cloud Firestore (Beta)
- Cloud Pub/Sub (Beta)
- Cloud Spanner (Beta)
- Cloud Speech (Beta)
- Cloud Video Intelligence (Beta)
- Cloud Vision (Beta)
- Google BigQuery (Beta)
- Google Stackdriver Monitoring (Beta)
This client supports the following Google Cloud Platform services at an Alpha quality level:
- Cloud Bigtable (Alpha)
- Cloud DNS (Alpha)
- Cloud Resource Manager (Alpha)
- Google Compute Engine (Alpha)
- Google Stackdriver Debugger (Alpha)
- Google Stackdriver Error Reporting (Alpha)
- Google Stackdriver Trace (Alpha)
The following client libraries are deprecated due to the underlying API also being deprecated:
- Google Prediction API (Deprecated)
If you need support for other Google APIs, check out the Google Node.js API Client library.
We recommend installing the individual packages that you need, which are provided under the @google-cloud
namespace. For example:
$ npm install --save @google-cloud/datastore
$ npm install --save @google-cloud/storage
var config = {
projectId: 'grape-spaceship-123',
keyFilename: '/path/to/keyfile.json'
};
var datastore = require('@google-cloud/datastore')(config);
var storage = require('@google-cloud/storage')(config);
We also provide a meta-package, google-cloud
, which provides all of the individual APIs. However, in order to keep file size and memory use low, the use of this package is not recommended.
If you want the kitchen sink, however, get it with:
$ npm install --save google-cloud
var gcloud = require('google-cloud')({
projectId: 'grape-spaceship-123',
keyFilename: '/path/to/keyfile.json'
});
var datastore = gcloud.datastore();
var storage = gcloud.storage();
- nodejs-getting-started - A sample and tutorial that demonstrates how to build a complete web application using Cloud Datastore, Cloud Storage, and Cloud Pub/Sub and deploy it to Google App Engine or Google Compute Engine.
- gcloud-node-todos - A TodoMVC backend using google-cloud-node and Datastore.
- gitnpm - Easily lookup an npm package's GitHub repo using google-cloud-node and Google App Engine.
- gcloud-kvstore - Use Datastore as a simple key-value store.
- hya-wave - Cloud-based web sample editor. Part of the hya-io family of products.
- gstore-node - Google Datastore Entities Modeling library.
- gstore-api - REST API builder for Google Datastore Entities.
With google-cloud
it's incredibly easy to get authenticated and start using Google's APIs. You can set your credentials on a global basis as well as on a per-API basis. See each individual API section below to see how you can auth on a per-API-basis. This is useful if you want to use different accounts for different Cloud services.
If you are running this client on Google Cloud Platform, we handle authentication for you with no configuration. You just need to make sure that when you set up the GCE instance, you add the correct scopes for the APIs you want to access.
var storage = require('@google-cloud/storage')();
// If you're using the google-cloud meta-package:
var gcloud = require('google-cloud');
var storage = gcloud.storage();
// ...you're good to go! See the next section to get started using the APIs.
If you are not running this client on Google Cloud Platform, you need a Google Developers service account. To create a service account:
- Visit the Google Developers Console.
- Create a new project or click on an existing project.
- Navigate to APIs & auth > APIs section and turn on the following APIs (you may need to enable billing in order to use these services):
- BigQuery API
- Cloud Bigtable API
- Cloud Bigtable Admin API
- Cloud Bigtable Table Admin API
- Cloud Spanner API
- Google Cloud Datastore API
- Google Cloud DNS API
- Google Cloud Firestore API
- Google Cloud Natural Language API
- Google Cloud Pub/Sub API
- Google Cloud Resource Manager API
- Google Cloud Speech API
- Google Cloud Storage
- Google Cloud Storage JSON API
- Google Cloud Translation API
- Google Cloud Vision API
- Google Compute Engine API
- Prediction API
- Stackdriver Logging API
- Navigate to APIs & auth > Credentials and then:
- If you want to use a new service account key, click on Create credentials and select Service account key. After the account key is created, you will be prompted to download the JSON key file that the library uses to authenticate your requests.
- If you want to generate a new service account key for an existing service account, click on Generate new JSON key and download the JSON key file.
// Authenticating on a global basis.
var projectId = process.env.GCLOUD_PROJECT; // E.g. 'grape-spaceship-123'
var gcloud = require('google-cloud')({
projectId: projectId,
// The path to your key file:
keyFilename: '/path/to/keyfile.json'
// Or the contents of the key file:
credentials: require('./path/to/keyfile.json')
// For any APIs that accept an API key:
key: '...'
});
// ...you're good to go! See the next section to get started using the APIs.
You can also set auth on a per-API-instance basis. The examples below show you how.
Follow the activation instructions to use the Cloud Datastore API with your project.
$ npm install --save @google-cloud/datastore
var datastore = require('@google-cloud/datastore');
See Authentication.
var datastoreClient = datastore({
projectId: 'grape-spaceship-123',
keyFilename: '/path/to/keyfile.json'
});
var key = datastoreClient.key(['Product', 'Computer']);
datastoreClient.get(key, function(err, entity) {
console.log(err || entity);
});
// Save data to Datastore.
var blogPostData = {
title: 'How to make the perfect homemade pasta',
author: 'Andrew Chilton',
isDraft: true
};
var blogPostKey = datastoreClient.key('BlogPost');
datastoreClient.save({
key: blogPostKey,
data: blogPostData
}, function(err) {
// `blogPostKey` has been updated with an ID so you can do more operations
// with it, such as an update.
blogPostData.isDraft = false;
datastoreClient.save({
key: blogPostKey,
data: blogPostData
}, function(err) {
if (!err) {
// The blog post is now published!
}
});
});
$ npm install --save @google-cloud/language
var language = require('@google-cloud/language');
See Authentication.
var languageClient = language({
projectId: 'grape-spaceship-123',
keyFilename: '/path/to/keyfile.json'
});
var content = 'Hello, world!';
var type = language.v1.types.Document.Type.PLAIN_TEXT;
var document = {
content : content,
type : type
};
languageClient.analyzeSentiment({document: document}).then(function(responses) {
var response = responses[0];
// doThingsWith(response)
})
.catch(function(err) {
console.error(err);
});
$ npm install --save @google-cloud/storage
var storage = require('@google-cloud/storage');
See Authentication.
var fs = require('fs');
var gcs = storage({
projectId: 'grape-spaceship-123',
keyFilename: '/path/to/keyfile.json'
});
// Create a new bucket.
gcs.createBucket('my-new-bucket', function(err, bucket) {
if (!err) {
// "my-new-bucket" was successfully created.
}
});
// Reference an existing bucket.
var bucket = gcs.bucket('my-existing-bucket');
// Upload a local file to a new file to be created in your bucket.
bucket.upload('/photos/zoo/zebra.jpg', function(err, file) {
if (!err) {
// "zebra.jpg" is now in your bucket.
}
});
// Download a file from your bucket.
bucket.file('giraffe.jpg').download({
destination: '/photos/zoo/giraffe.jpg'
}, function(err) {});
// Streams are also supported for reading and writing files.
var remoteReadStream = bucket.file('giraffe.jpg').createReadStream();
var localWriteStream = fs.createWriteStream('/photos/zoo/giraffe.jpg');
remoteReadStream.pipe(localWriteStream);
var localReadStream = fs.createReadStream('/photos/zoo/zebra.jpg');
var remoteWriteStream = bucket.file('zebra.jpg').createWriteStream();
localReadStream.pipe(remoteWriteStream);
$ npm install --save @google-cloud/translate
var translate = require('@google-cloud/translate');
See Authentication.
var translateClient = translate({
projectId: 'grape-spaceship-123',
keyFilename: '/path/to/keyfile.json'
});
// Translate a string of text.
translateClient.translate('Hello', 'es', function(err, translation) {
if (!err) {
// translation = 'Hola'
}
});
// Detect a language from a string of text.
translateClient.detect('Hello', function(err, results) {
if (!err) {
// results = {
// language: 'en',
// confidence: 1,
// input: 'Hello'
// }
}
});
// Get a list of supported languages.
translateClient.getLanguages(function(err, languages) {
if (!err) {
// languages = [
// 'af',
// 'ar',
// 'az',
// ...
// ]
}
});
$ npm install --save @google-cloud/logging
var logging = require('@google-cloud/logging');
See Authentication.
var loggingClient = logging({
projectId: 'grape-spaceship-123',
keyFilename: '/path/to/keyfile.json'
});
// Create a sink using a Bucket as a destination.
var gcs = storage();
loggingClient.createSink('my-new-sink', {
destination: gcs.bucket('my-sink')
}, function(err, sink) {});
// Write a critical entry to a log.
var syslog = loggingClient.log('syslog');
var metadata = {
resource: {
type: 'gce_instance',
labels: {
zone: 'global',
instance_id: '3'
}
}
};
var entry = syslog.entry(metadata, {
delegate: process.env.user
});
syslog.critical(entry, function(err) {});
// Get all entries in your project.
loggingClient.getEntries(function(err, entries) {
if (!err) {
// `entries` contains all of the entries from the logs in your project.
}
});
$ npm install --save @google-cloud/firestore
const Firestore = require('@google-cloud/firestore');
See Authentication.
const firestore = new Firestore({
projectId: 'YOUR_PROJECT_ID',
keyFilename: '/path/to/keyfile.json',
});
const document = firestore.doc('posts/intro-to-firestore');
// Enter new data into the document.
document.set({
title: 'Welcome to Firestore',
body: 'Hello World',
}).then(() => {
// Document created successfully.
});
// Update an existing document.
document.update({
body: 'My first Firestore app',
}).then(() => {
// Document updated successfully.
});
// Read the document.
document.get().then(doc => {
// Document read successfully.
});
// Delete the document.
document.delete().then(() => {
// Document deleted successfully.
});
$ npm install --save @google-cloud/pubsub
var pubsub = require('@google-cloud/pubsub');
See Authentication.
var pubsubClient = pubsub({
projectId: 'grape-spaceship-123',
keyFilename: '/path/to/keyfile.json'
});
// Reference a topic that has been previously created.
var topic = pubsubClient.topic('my-topic');
// Publish a message to the topic.
var publisher = topic.publisher();
var message = new Buffer('New message!');
publisher.publish(message, function(err, messageId) {});
// Subscribe to the topic.
topic.createSubscription('subscription-name', function(err, subscription) {
// Register listeners to start pulling for messages.
function onError(err) {}
function onMessage(message) {}
subscription.on('error', onError);
subscription.on('message', onMessage);
// Remove listeners to stop pulling for messages.
subscription.removeListener('message', onMessage);
subscription.removeListener('error', onError);
});
$ npm install --save @google-cloud/spanner
var spanner = require('@google-cloud/spanner');
See Authentication.
var spannerClient = spanner({
projectId: 'grape-spaceship-123',
keyFilename: '/path/to/keyfile.json'
});
var instance = spannerClient.instance('my-instance');
var database = instance.database('my-database');
// Create a table.
var schema = `
CREATE TABLE Singers (
SingerId INT64 NOT NULL,
FirstName STRING(1024),
LastName STRING(1024),
SingerInfo BYTES(MAX),
) PRIMARY KEY(SingerId)
`;
database.createTable(schema, function(err, table, operation) {
if (err) {
// Error handling omitted.
}
operation
.on('error', function(err) {})
.on('complete', function() {
// Table created successfully.
});
});
// Insert data into the table.
var table = database.table('Singers');
table.insert({
SingerId: 10,
FirstName: 'Eddie',
LastName: 'Wilson'
}, function(err) {
if (!err) {
// Row inserted successfully.
}
});
// Run a query as a readable object stream.
database.runStream('SELECT * FROM Singers')
.on('error', function(err) {})
.on('data', function(row) {
// row.toJSON() = {
// SingerId: 10,
// FirstName: 'Eddie',
// LastName: 'Wilson'
// }
}
})
.on('end', function() {
// All results retrieved.
});
$ npm install --save @google-cloud/speech
var speech = require('@google-cloud/speech');
See Authentication.
var speechClient = speech({
projectId: 'my-project',
keyFilename: '/path/to/keyfile.json'
});
var languageCode = 'en-US';
var sampleRateHertz = 44100;
var encoding = speech.v1.types.RecognitionConfig.AudioEncoding.FLAC;
var config = {
languageCode : languageCode,
sampleRateHertz : sampleRateHertz,
encoding : encoding
};
var uri = 'gs://gapic-toolkit/hello.flac';
var audio = {
uri : uri
};
var request = {
config: config,
audio: audio
};
speechClient.recognize(request).then(function(responses) {
var response = responses[0];
// doThingsWith(response)
})
.catch(function(err) {
console.error(err);
});
$ npm install --save @google-cloud/vision
var vision = require('@google-cloud/vision');
See Authentication.
var visionClient = vision({
projectId: 'grape-spaceship-123',
keyFilename: '/path/to/keyfile.json'
});
var gcsImageUri = 'gs://gapic-toolkit/President_Barack_Obama.jpg';
var source = {
gcsImageUri : gcsImageUri
};
var image = {
source : source
};
var type = vision.v1.types.Feature.Type.FACE_DETECTION;
var featuresElement = {
type : type
};
var features = [featuresElement];
var requestsElement = {
image : image,
features : features
};
var requests = [requestsElement];
visionClient.batchAnnotateImages({requests: requests}).then(function(responses) {
var response = responses[0];
// doThingsWith(response)
})
.catch(function(err) {
console.error(err);
});
$ npm install --save @google-cloud/bigquery
var bigquery = require('@google-cloud/bigquery');
See Authentication.
var bigqueryClient = bigquery({
projectId: 'grape-spaceship-123',
keyFilename: '/path/to/keyfile.json'
});
// Access an existing dataset and table.
var schoolsDataset = bigqueryClient.dataset('schools');
var schoolsTable = schoolsDataset.table('schoolsData');
// Import data into a table.
schoolsTable.import('/local/file.json', function(err, job) {});
// Get results from a query job.
var job = bigqueryClient.job('job-id');
// Use a callback.
job.getQueryResults(function(err, rows) {});
// Or get the same results as a readable stream.
job.getQueryResults().on('data', function(row) {});
It does not follow the conventions you're familiar with from other parts of our library. A handwritten layer is not yet available.
The example below shows you how to instantiate the generated client. For further documentation, please browse the Monitoring .proto files on GitHub.
$ npm install --save @google-cloud/monitoring
var monitoring = require('@google-cloud/monitoring');
See Authentication.
var client = monitoring.metric({
// optional auth parameters.
});
// Iterate over all elements.
var formattedName = client.projectPath(projectId);
client.listMonitoredResourceDescriptors({name: formattedName}).then(function(responses) {
var resources = responses[0];
for (var i = 0; i < resources.length; ++i) {
// doThingsWith(resources[i])
}
})
.catch(function(err) {
console.error(err);
});
// Or obtain the paged response.
var formattedName = client.projectPath(projectId);
var options = {autoPaginate: false};
function callback(responses) {
// The actual resources in a response.
var resources = responses[0];
// The next request if the response shows there's more responses.
var nextRequest = responses[1];
// The actual response object, if necessary.
// var rawResponse = responses[2];
for (var i = 0; i < resources.length; ++i) {
// doThingsWith(resources[i]);
}
if (nextRequest) {
// Fetch the next page.
return client.listMonitoredResourceDescriptors(nextRequest, options).then(callback);
}
}
client.listMonitoredResourceDescriptors({name: formattedName}, options)
.then(callback)
.catch(function(err) {
console.error(err);
});
You may need to create an instance to use the Cloud Bigtable API with your project.
$ npm install --save @google-cloud/bigtable
var bigtable = require('@google-cloud/bigtable');
See Authentication.
var bigtableClient = bigtable({
projectId: 'grape-spaceship-123',
keyFilename: '/path/to/keyfile.json'
});
var instance = bigtableClient.instance('my-instance');
var table = instance.table('prezzy');
table.getRows(function(err, rows) {});
// Update a row in your table.
var row = table.row('alincoln');
row.save('follows:gwashington', 1, function(err) {
if (err) {
// Error handling omitted.
}
row.get('follows:gwashington', function(err, data) {
if (err) {
// Error handling omitted.
}
// data = {
// follows: {
// gwashington: [
// {
// value: 1
// }
// ]
// }
// }
});
});
$ npm install --save @google-cloud/dns
var dns = require('@google-cloud/dns');
See Authentication.
var dnsClient = dns({
projectId: 'grape-spaceship-123',
keyFilename: '/path/to/keyfile.json'
});
// Create a managed zone.
dnsClient.createZone('my-new-zone', {
dnsName: 'my-domain.com.'
}, function(err, zone) {});
// Reference an existing zone.
var zone = dnsClient.zone('my-existing-zone');
// Create an NS record.
var nsRecord = zone.record('ns', {
ttl: 86400,
name: 'my-domain.com.',
data: 'ns-cloud1.googledomains.com.'
});
zone.addRecords([nsRecord], function(err, change) {});
// Create a zonefile from the records in your zone.
zone.export('/zonefile.zone', function(err) {});
$ npm install --save @google-cloud/resource
var resource = require('@google-cloud/resource');
See Authentication.
var resourceClient = resource({
projectId: 'grape-spaceship-123',
keyFilename: '/path/to/keyfile.json'
});
// Get all of the projects you maintain.
resourceClient.getProjects(function(err, projects) {
if (!err) {
// `projects` contains all of your projects.
}
});
// Get the metadata from your project. (defaults to `grape-spaceship-123`)
var project = resourceClient.project();
project.getMetadata(function(err, metadata) {
// `metadata` describes your project.
});
$ npm install --save @google-cloud/compute
var compute = require('@google-cloud/compute');
See Authentication.
var gce = compute({
projectId: 'grape-spaceship-123',
keyFilename: '/path/to/keyfile.json'
});
// Create a new VM using the latest OS image of your choice.
var zone = gce.zone('us-central1-a');
var name = 'ubuntu-http';
zone.createVM(name, { os: 'ubuntu' }, function(err, vm, operation) {
// `operation` lets you check the status of long-running tasks.
operation
.on('error', function(err) {})
.on('running', function(metadata) {})
.on('complete', function(metadata) {
// Virtual machine created!
});
});
$ npm install --save @google-cloud/prediction
var prediction = require('@google-cloud/prediction');
See Authentication.
var predictionClient = prediction({
projectId: 'grape-spaceship-123',
keyFilename: '/path/to/keyfile.json'
});
// Get all of the trained models in your project.
predictionClient.getModels(function(err, models) {
if (!err) {
// `models` is an array of Model objects.
}
});
// Reference an existing trained model.
var model = predictionClient.model('my-existing-model');
// Train a model.
model.train('english', 'Hello from your friends at Google!', function(err) {});
// Query a model.
model.query('Hello', function(err, results) {
if (!err) {
// results.winner == 'english'
// results.scores == [
// {
// label: 'english',
// score: 1
// },
// {
// label: 'spanish',
// score: 0
// }
// ]
}
});
The source code for the Node.js Cloud Debugger Agent lives in a separate repo.
$ npm install --save @google-cloud/debug-agent
require('@google-cloud/debug-agent').start({ allowExpressions: true });
For more details on API usage, please see the Stackdriver Debug Agent Github Repository.
$ npm install --save @google-cloud/error-reporting
The module provides automatic uncaught exception handling, manual error reporting, and integration with common frameworks like express and hapi.
var errors = require('@google-cloud/error-reporting')();
See Authentication.
errors.report(new Error('Something broke!'));
For more details on API usage, please see the documentation.
The source code for the Node.js Cloud Trace Agent lives in a separate repo.
$ npm install --save @google-cloud/trace-agent
var trace = require('@google-cloud/trace-agent').start();
For more details on API usage, please see the Stackdriver Trace Agent Github Repository.
This library follows Semantic Versioning.
Please note it is currently under active development. Any release versioned 0.x.y
is subject to backwards-incompatible changes at any time.
GA: Libraries defined at the GA (general availability) quality level are stable. The code surface will not change in backwards-incompatible ways unless absolutely necessary (e.g. because of critical security issues) or with an extensive deprecation period. Issues and requests against GA libraries are addressed with the highest priority.
Please note that the auto-generated portions of the GA libraries (the ones in modules such as v1
or v2
) are considered to be of Beta quality, even if the libraries that wrap them are GA.
Beta: Libraries defined at the Beta quality level are expected to be mostly stable, while we work towards their release candidate. We will address issues and requests with a higher priority.
Alpha: Libraries defined at the Alpha quality level are still a work-in-progress and are more likely to get backwards-incompatible updates.
Contributions to this library are always welcome and highly encouraged.
See CONTRIBUTING for more information on how to get started.
Apache 2.0 - See LICENSE for more information.