Here we describe how to do auth with a package that uses gargle, without requiring any user interaction. This comes up in a wide array of contexts, ranging from simple rendering of a local R Markdown document to deploying a data product on a remote server.
We assume the wrapper package uses the design described in
vignette("gargle-auth-in-client-package")
. Examples
include:
Full details on gargle::token_fetch()
, which powers this
strategy, are given in
vignette("how-gargle-gets-tokens")
.
In certain cloud computing contexts, a service account token may be ambiently available (or you can arrange for that to be true). Think about it: if your workload is running on Google Compute Engine (GCE), it’s already “inside the Google house”. It seems like there should be a way to avoid another round of auth and that is indeed the case.
Another advantage of these cloud auth workflows is that there is never any need to download and carefully manage a file that contains sensitive information. This is why they are often described as “keyless”. If you can use one of these methods, you should seriously consider doing so.
This section applies to code running on a GCE instance, either literally, or on another Google Cloud product built on top of GCE. You should consider Google’s own documentation to be definitive, but we’ll try to give a useful summary here and to explain how gargle works with GCE:
https://cloud.google.com/compute/docs/access/create-enable-service-accounts-for-instances
A Google Cloud Platform (GCP) project generally has a GCE default service account and, by default, a new GCE instance runs as that service account. (If you wish, you can use a different service account by taking explicit steps when you create an instance or by modifying it while it’s stopped.) The main point is that, for an application running on GCE, a service account identity is generally available.
GCE allows applications to get an OAuth access token from its
metadata server and this is what gargle::credentials_gce()
does (which is one of functions tried by
gargle::token_fetch()
, which is called by wrapper
packages). This token request can be made for specific scopes and, in
general, most wrapper packages will indeed be asking for specific scopes
relevant to the API they access. Consider the signature of
googledrive::drive_auth()
:
drive_auth <- function(email = gargle::gargle_oauth_email(),
path = NULL,
scopes = "https://www.googleapis.com/auth/drive",
cache = gargle::gargle_oauth_cache(),
use_oob = gargle::gargle_oob_default(),
token = NULL) { ... }
The googledrive package asks for a token with "drive"
scope, by default. This brings up one big gotcha when using packages
like googledrive or googlesheets4 on GCE.
By default, a GCE instance will be running as the default service
account, with the "cloud-platform"
scope and this will,
generally speaking, allow the service account to work with various Cloud
products. However, the "cloud-platform"
scope does not
permit operations with non-Cloud APIs, such as Drive and Sheets. If you
want the service account identity for your GCE instance to be able to
get an access token for use with Drive and Sheets, you will need to
explicitly add, e.g., the "drive"
scope when you create the
instance (or stop the instance and add that scope). (Note that, in
contrast, BigQuery is considered a Cloud product and therefore bigrquery
can operate with the "cloud-platform"
scope.)
Be aware that you might also need to explicitly grant the service account an appropriate level of access (e.g. read or write) to any Drive files you intend to work on.
Finally, if you want to opt-out of using the default service account
and, instead, auth as a normal user, even though you are on GCE, that is
also possible. One way to achieve that is to remove
credentials_gce()
from the set of auth functions tried by
gargle::token_fetch()
by executing this command before any
explicit or implicit auth happens:
You can make a similar change in more scoped way with the helpers
gargle::with_cred_funs()
or
gargle::local_cred_funs()
.
Here we discuss how gargle’s GCE auth can work for a related service, Google Kubernetes Engine (GKE), using Workload Identity. This is more complicated that direct usage of GCE and some extra configuration is needed to make a service account’s metadata available for the GKE instance to discover. GKE is the underlying technology behind Google’s managed Airflow service, Cloud Composer, so this also applies to R docker files being called in that environment.
Workload Identity is the recommended way to do authentication on GKE and other places, if possible, since it eliminates the use of a file that holds the service key, which is a potential security risk.
my-service-key@my-project.iam.gserviceaccount.com
has the
https://www.googleapis.com/auth/bigquery
scope.# create namespace
kubectl create namespace my-namespace
# Create Kubernetes service account
kubectl create serviceaccount --namespace my-namespace bq-service-account
# Create IAM policy binding betwwen k8s SA and GSA
gcloud iam service-accounts add-iam-policy-binding my-service-key@my-project.iam.gserviceaccount.com \
--role roles/iam.workloadIdentityUser \
--member "serviceAccount:my-project.svc.id.goog[my-namespace/bq-service-account]"
# Annotate k8s SA
kubectl annotate serviceaccount bq-service-account \
--namespace my-namespace \
iam.gke.io/gcp-service-account=my-service-key@my-project.iam.gserviceaccount.com
This key will now be available to add to pods within the cluster. For
Airflow, you can pass them in using the Python code
GKEStartPodOperator(...., namespace='my-namespace', service_account_name='bq-service-account')
.
Documentation around GKEStartPodOperator()
within Cloud
Composer can be found here.
gargle::gce_credentials()
do the right thing, you need to do two things:"gargle.gce.use_ip"
option to TRUE
, in
order to use the metadata server that’s relevant on GKE."default"
service
account. gce_instance_service_accounts()
can be helpful,
e.g., if you want to know which service accounts your Docker container
can see.Here is example code that you might execute in your Docker container:
options(gargle.gce.use_ip = TRUE)
t <- gargle::credentials_gce("my-service-key@my-project.iam.gserviceaccount.com")
# ... do authenticated stuff with the token t ...
Let’s assume that PKG is an R package that implements gargle auth in
the standard way, such as bigrquery or googledrive. At the time of
writing the service_account
argument is not exposed in the
usual, high-level PKG_auth()
function (https://github.com/r-lib/gargle/issues/249. So if you
need to use a non-default
service account, you need to call
credentials_gce()
directly and pass that token to
PKG_auth()
: Here’s an example of how that might look:
Keyless auth is even possible from non-Google cloud platforms, using Workload identity federation.
This is implemented in the experimental function
credentials_external_account()
, which currently only
supports AWS.
When two computers are talking to each other, possibly with no human involvement, the most appropriate type of token to use is a service account token. If you’re not working in cloud context with automatic access to a service account (see previous section), you can still use a service account, but it will require more explicit effort.
vignette("get-api-credentials")
,
specifically in the Service account token section.Example using googledrive:
If this code is running on, e.g., a continuous integration service
and you need to use an encrypted token, see
vignette("managing-tokens-securely")
.
For certain APIs, service accounts are inherently awkward, because
you often want to do things on behalf of a specific user. Gmail
is a good example. If you are sending email programmatically, you
probably want to send it as yourself (or from some other specific email
account) instead of from
zestybus-geosyogl@fuffapster-654321.iam.gserviceaccount.com
.
This is, in fact, possible but is described as “impersonation”, which
should tip you off that Google does not exactly encourage this workflow.
Some details:
subject
argument of
credentials_service_account()
and
credentials_app_default()
is available to specify which
user to impersonate, e.g. subject = "user@example.com"
.
This argument first appeared in gargle 0.5.0, so it may not necessarily
be exposed yet in user-facing auth functions like
drive_auth()
. If you need subject
in a client
package, that is a reasonable feature request. It is also possible to
get a token with an explicit call to, e.g.,
credentials_service_account()
and then pass that token to
the auth function:t <- gargle::credentials_service_account(
path = "/path/to/your/service-account-token.json",
scopes = ...,
subject = "user@example.com"
)
googledrive::dive_auth(token = t)
If delegation of domain-wide authority is impossible or unappealing, you must use an OAuth user token, as described below.
Wrapper packages that use gargle::token_fetch()
in the
recommended way have access to the token search strategy known as
Application Default Credentials.
You need to put the JSON corresponding to your service or external account in a very specific location or, alternatively, record the location of this JSON file in a specific environment variable.
Full details are in the credentials_app_default()
section of vignette("how-gargle-gets-tokens")
.
If you have your token rigged properly, you do not
need to do anything else, i.e. you do not need to call
PACKAGE_auth()
explicitly. Your token should just get
discovered upon first need.
For troubleshooting purposes, you can set a gargle option to see
verbose output about the execution of
gargle::token_fetch()
:
withr-style convenience helpers also exist:
with_gargle_verbosity()
and
local_gargle_verbosity()
.
If you somehow have the OAuth token you want to use as an R object,
you can provide it directly to the token
argument of the
main auth function. Example using googledrive:
library(googledrive)
my_oauth_token <- # some process that results in the token you want to use
drive_auth(token = my_oauth_token)
gargle caches each OAuth user token it obtains to an
.rds
file, by default. If you know the filepath to the
token you want to use, you could use readRDS()
to read it
and provide as the token
argument to the wrapper’s auth
function. Example using googledrive:
How would you know this filepath? That requires some attention to the location of gargle’s OAuth token cache folder, which is described in the next section.
Full details are in the credentials_byo_oauth2()
section
of vignette("how-gargle-gets-tokens")
.
This is the least recommended strategy, but it appeals to many users, because it doesn’t require creating a service account. Just remember that the perceived ease of using the token you already have (an OAuth user token) is quickly cancelled out by the greater difficulty of managing such tokens for non-interactive use. You might be forced to use this strategy with certain APIs, such as Gmail, that are difficult to use with a service account.
Two main principles:
There are many ways to do this. We’ll work several examples using that convey the range of what’s possible.
.Rmd
to renderStep 1: Get that first token. You must run your code at least once, interactively, do the auth dance, and allow gargle to store the token in its cache.
Step 2: Revise your code to pre-authorize the use of
that token next time. Now your .Rmd
can be rendered or your
.R
script can run, without further interaction.
You have two choices to make:
gargle_oauth_email
option or call
PACKAGE_auth(email = ...)
.
.Rmd
or .R
or in a user-level or project level
.Rprofile
startup file.email = TRUE
works if we’re only going to find, at
most, 1 token, i.e. you always auth with the same identityemail = "jane@example.com"
pre-authorizes use of a
token associated with a specific identityemail = "*@example.com"
pre-authorizes use of a token
associated with an identity from a specific domain; good for code that
might be executed on the machines of both alice@example.com
and bob@example.com
This sets an option that allows gargle to use cached tokens whenever there’s a unique match:
This sets an option to use tokens associated with a specific email address:
This sets an option to use tokens associated with an email address with a specific domain:
This gets a token right now and allows the use of a matching token, using googledrive as an example:
This gets a token right now, for the user with a specific email address:
This gets a token right now, first checking the cache for a token associated with a specific domain:
This is like the previous example, but with an added twist: we use a project-level OAuth cache. This is good for deployed data products.
Step 1: Obtain the token intended for non-interactive use and make sure it’s cached in a (hidden) directory of the current project. Using googledrive as an example:
library(googledrive)
# designate project-specific cache
options(gargle_oauth_cache = ".secrets")
# check the value of the option, if you like
gargle::gargle_oauth_cache()
# trigger auth on purpose --> store a token in the specified cache
drive_auth()
# see your token file in the cache, if you like
list.files(".secrets/")
Do this setup once per project.
Another way to accomplish the same setup is to specify the desired cache location directly in the call to the auth function:
library(googledrive)
# trigger auth on purpose --> store a token in the specified cache
drive_auth(cache = ".secrets")
Step 2: In all downstream use, announce the location of the cache and pre-authorize the use of a suitable token discovered there. Continuing the googledrive example:
library(googledrive)
options(
gargle_oauth_cache = ".secrets",
gargle_oauth_email = TRUE
)
# now use googledrive with no need for explicit auth
drive_find(n_max = 5)
Setting the option gargle_oauth_email = TRUE
says that
googledrive is allowed to use a token that it finds in the cache,
without interacting with a user, as long as it discovers EXACTLY one
matching token. This option-setting code needs to appear in each script,
.Rmd
, or app that needs to use this token
non-interactively. Depending on the context, it might be suitable to
accomplish this in a startup file, e.g. project-level
.Rprofile
.
Here’s a variation where we say which token to use by explicitly specifying the associated email. This is handy if there’s a reason to have more than one token in the cache.
library(googledrive)
options(
gargle_oauth_cache = ".secrets",
gargle_oauth_email = "jenny@example.com"
)
# now use googledrive with no need for explicit auth
drive_find(n_max = 5)
Here’s another variation where we specify the necessary info directly in an auth call, instead of in options:
library(googledrive)
drive_auth(cache = ".secrets", email = TRUE)
# now use googledrive with no need for explicit auth
drive_find(n_max = 5)
Here’s one last variation that’s applicable when the local cache could contain multiple tokens:
library(googledrive)
drive_auth(cache = ".secrets", email = "jenny@example.com")
# now use googledrive with no need for explicit auth
drive_find(n_max = 5)
Be very intentional about paths and working directory. Personally I
would use here::here(".secrets)"
everywhere above, to make
things more robust.
For troubleshooting purposes, you can set a gargle option to see
verbose output about the execution of
gargle::token_fetch()
:
withr-style convenience helpers also exist:
with_gargle_verbosity()
and
local_gargle_verbosity()
.
For a cached token to be considered a “match”, it must match the current request with respect to user’s email, scopes, and OAuth client (client ID or key and secret). By design, these settings have very low visibility, because we usually want to use the defaults. If your token is not being discovered, consider if any of these fields might explain the mismatch.