The Wavefront console is excellent, and its API coverage is complete and simple to use, but there are many use-cases where even a halfway-decent CLI will be better than the finest UI or the simplest API.
With that in mind, I tried to write a halfway-decent Wavefront CLI.
This article is updated whenever the CLI acquires new features.
Installation and Basics
My CLI is written in Ruby. To use it you need a current Ruby installation. At the time of writing, that’s 2.4 or later. (Versions prior to 5.0 supported Ruby 2.3.)
$ gem install wavefront-cli
A lot of care has been taken to ensure there are no “native extension” gems anywhere in the chain, so installation should be quick and painless on any host. I hate people thinking it’s fine to expect me to install a C compiler to run a hundred-line tool written in a scripting language.
Following the model of the best designed CLI I know, there’s a single command, with subcommands.
$ wf --help Wavefront CLI Usage: wf command [options] wf --version wf --help Commands: account view and manage Wavefront accounts alert view and manage alerts apitoken view and your own API tokens cloudintegration view and manage cloud integrations config create and manage local configuration dashboard view and manage dashboards derivedmetric view and manage derived metrics event open, close, view, and manage events ingestionpolicy view and manage ingestion policies integration view and manage Wavefront integrations link view and manage external links message read and mark user messages metric get metric details notificant view and manage Wavefront alert targets proxy view and manage proxies query run Wavefront queries role view and manage roles savedsearch view and manage saved searches serviceaccount view and manage service accounts settings view and manage system preferences source view and manage source tags and descriptions spy monitor traffic going into Wavefront usage view and manage usage reports usergroup view and manage Wavefront user groups webhook view and manage webhooks window view and manage maintenance windows write send data to Wavefront Use 'wf <command> --help' for further information.
The majority of those commands talk to Wavefront’s API. To do that,
wf obviously needs to know where the API is, and to pass on some
You can do this with command-line options (we’ll see those in a moment), but for everyday interactive use, it’s much better to create a configuration. file.
You can, of course, create configuration by hand, but the
config command will guide you through it. The first time you run
the program without credentials, it tells you how to create
configuration, and even suggests helpful default values for most
$ wf alert list No credentials supplied on the command line or via environment variables, and no configuration file found. Please run 'wf config setup' to create configuration. $ wf config setup Creating new configuration file at /home/rob/.wavefront. Creating profile 'default'. Wavefront API token:> 820ac1de-4e1f-41a4-f9c3-231c95ae4da1↵ Wavefront API endpoint [metrics.wavefront.com]:>↵ Wavefront proxy endpoint [wavefront]:> wavefront.localnet↵ default output format [human]:>↵ $ wf config show [default] token = 820ac1de-4e1f-41a4-f9c3-231c95ae4da1 endpoint = metrics.wavefront.com proxy = wavefront.localnet format = human
You can override values in the configuration files with command-line
options, and also with environment variables.
WAVEFRONT_PROXY are all supported. If
you have multiple Wavefront accounts, you can add a new stanza for
Now we’re fully credentialled, we can start exploring the CLI. Let’s start with some alerts. Wavefront is great at alerts.
$ wf alert --help Usage: wf alert list [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] [-al] [-O fields] [-o offset] [-L limit] wf alert firing [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] [-o offset] [-L limit] wf alert affected [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] hosts [<id>] wf alert snoozed [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] [-o offset] [-L limit] wf alert describe [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] [-v version] <id> wf alert delete [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] <id> wf alert clone [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] [-v version] <id> wf alert undelete [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] <id> wf alert history [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] [-o offset] [-L limit] <id> wf alert latest [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] <id> wf alert dump [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] wf alert import [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] [-uU] <file> wf alert snooze [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] [-T time] <id> wf alert set [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] <key=value> <id> wf alert unsnooze [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] <id> wf alert search [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] [-al] [-o offset] [-L limit] [-O fields] <condition>... wf alert tags [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] <id> wf alert tag set [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] <id> <tag>... wf alert tag clear [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] <id> wf alert tag add [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] <id> <tag> wf alert tag delete [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] <id> <tag> wf alert tag pathsearch [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] [-al] [-o offset] [-L limit] <word> wf alert currently [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] <state> wf alert queries [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] [-b] [<id>] wf alert install [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] <id> wf alert uninstall [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] <id> wf alert acls [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] <id> wf alert acl [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] clear <id> wf alert acl [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] grant (view | modify) on <id> to <name>... wf alert acl [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] revoke (view | modify) on <id> from <name>... wf alert summary [-DnVM] [-c file] [-P profile] [-E endpoint] [-t token] [-f format] [-a] wf alert --help Global options: -c, --config=FILE path to configuration file -P, --profile=NAME profile in configuration file -D, --debug enable debug mode -n, --noop do not perform API calls -V, --verbose be verbose -f, --format=STRING output format -M, --items-only only show items in machine-parseable formats -h, --help show this message Options: -E, --endpoint=URI cluster endpoint -t, --token=TOKEN Wavefront authentication token -l, --long list alerts in detail -a, --all list all alerts -v, --version=INTEGER describe only this version of alert -o, --offset=n start from nth alert -L, --limit=COUNT number of alerts to list -O, --fields=F1,F2,... only show given fields -u, --update update an existing alert -U, --upsert import new or update existing alert -T, --time=SECONDS how long to snooze (default 3600) -b, --brief do not show alert names
Notice the line-wrapping on the help: it automatically adjusts to fit the width of your terminal, and I’m an unapologetic, hardcore, 80-column guy. Deal with it.
Let’s start by having a look at the alerts in my account.
$ wf alert list 1459508340708 CHECKING Point Rate 1463413550083 CHECKING JPC Failed Services ...
Pretty much every command has a
list subcommand, and it will give
you a one-item-per-line listing by default, where the first column
is the unique identifer of the resource. Despite what I
said earlier about wrapping lines to fit the terminal, brief
listings don’t do that. That’s so you can always trust a command
wf proxy list | wc -l to give the answer you expect.
Every command has a sensible set of fields it will
list, but you
-O to give a comma-separated list of your own, should you
$ wf alert list -O id,additionalInformation 1459508340708 Fires if we exceed our agreed point rate 1463413550083 A service has failed. Log on to the box and see what it is ...
You can also
list -l, which more-or-less dumps all of every
resource into your terminal. I don’t often use that. Using
-l gives you only the fields you request, one
per-line. Items are separated by a blank line.
Pagination: offset, limits, and “all”
search will return the first hundred objects
they find. You will be informed if there are more objects, and can
--limit flags (
-L) to get the
search and almost all
list commands take a
-a) option, which fetches all objects of the given type. This
can be a heavy operation, if you have a lot of large objects,
source objects is such hard work that the
wf source list
command does not accept
--all: it just takes for ever.
We can also
search for alerts, or, indeed, for any other object
type. (All commands support the
search sub-command, so long as
their ultimate API endpoint supports it.)
When searching you can define multiple conditions, which the
Wavefront engine will
AND together to refine a query. Conditions
are specified as
key=value. Or, if you wish to search for objects
key field merely contains
If you want objects where the field starts with the value, use
key^value. The default display mode for
search subcommands is
one object per line, and the fields will be the object’s
whichever other keys you used in your conditions. You can negate
conditions by putting a
! in front of the search operator.
$ wf alert search name~JPC 1497275466684 JPC Failed Services 1463413760189 JPC Memory Shortage 1490980663852 JPC: no metrics $ wf alert search name~JPC name!~Memory 1497275466684 JPC Failed Services 1490980663852 JPC: no metrics $ wf alert search name~JPC id^149 1497275466684 JPC Failed Services 1490980663852 JPC: no metrics $ wf alert search name~JPC id^149 severity=SMOKE 1497275466684 JPC Failed Services SMOKE $ wf alert search status=SNOOZED 1481553823153 SNOOZED $ wf alert search status=SNOOZED name~' ' 1481553823153 SNOOZED JVM Memory $ wf alert search status=SNOOZED name~' ' 1481553823153 SNOOZED JVM Memory
search with the
--long) flag will show you the
entire matching object. Using a machine-parseable output format also
returns the whole of the matching object.
Wavefront gives you a couple of “magic” search keys:
freetext. Object tags are a structure, and the
tags search key
looks across all tags.
wf tries to present this potentially
multi-dimensional data in a simple way.
$ wf alert search tags=physical 1534951532204 customerTags=backup,home,physical 1499780986548 customerTags=disk,physical,storage 1476741941156 customerTags=disk,physical
Freetext searches look at every field in an alert, so can potentially
return a lot of data. If you run a freetext search without
you’ll get a list of matching objects paired with a list of the
fields which matched your pattern.
$ wf alert search freetext=ZFS 1499780986548 name, event 1489162558204 additionalInformation
created 2016-05-16 15:45:50.083 minutes 2 name JPC Failed Services id 1463413550083 target firstname.lastname@example.org, tags customerTags JPC status CHECKING inTrash false updateUserId email@example.com lastProcessedMillis 2017-06-12 10:58:30.534 pointsScannedAtLastQuery 0 createdEpochMillis 2016-05-16 15:45:50.083 updatedEpochMillis 2016-05-16 15:50:08.168 updaterId firstname.lastname@example.org condition ts("dev.diamond.host.smf.svcs.maintenance", host="74a247a9-f67c-43ad-911f-fabafa9dc2f3") > 0 updated 2016-05-16 15:50:08.168 severity SMOKE additionalInformation An SMF service has failed. Log on to the box and see what it is. deleted false
The data in a
describe command is usually massaged.
The top-level time-related values have been changed from epoch
milliseconds to a more human-readable format. Also, some data which
is read-only and very unlikely to be useful has been omitted for the
sake of clarity. By default the CLI prints its results in a “human
readable”format, which may not always be what you want. So, we offer
three other format, all selectable with the
-f option. They are
ruby. The first two should be self-explanatory,
ruby dumps a string of the raw Ruby object from which all the
other output formats are constructed. It could be useful for pasting
irb, or generating test data.
Returning to the output above, we see a failing service is only SMOKE? That can’t be right, surely. Let’s fix it.
$ wf alert set severity=SEVERE 1463413550083 | grep severity severity SEVERE
I used that
grep because setting a value in an object will
re-display said object with its new values, and I didn’t want to
show you the whole lot again. It only updated the
me. Be aware lots of poperties are read-only, at least via the API.
Deleting and Undeleting
Actually, you know what? I changed my mind. I don’t care if a service fails on a box. I monitor my application, not boxes. If the application is up and latency is acceptable, that’s all I care about. Let’s get rid of that alert.
$ wf alert delete 1463413550083 Soft deleting alert '1463413550083'. Deleted alert '1463413550083'. $ wf alert describe 1463413550083 API 404: Alert 1463413550083 does not exist.
Thinking about it, knowing whether or not a service stopped could make debugging an outage an awful lot simpler. Fortunately it’s only “soft deleted”, which means it can be got back
$ wf alert undelete 1463413550083 Undeleted alert '1463413550083'.
Histories and Revisions
Remember when we modified the alert earlier? Wavefront does.
$ wf alert history 1463413550083 -L1 id 1463413550083 inTrash false version 5 updateUser email@example.com updateTime 1497273637816 changeDescription Alert severity updated from SMOKE to SEVERE
-L1 specifies that we only want to see the last revision to
the alert. Without it you’d get the entire history. You see the
version number? You use that with the
describe command we saw
earlier to get a past alert definition. Clearly version 5 introduced
SEVERE change, so version 4 should have a severity
SMOKE. Instead of
grepping, let’s use JSON output and parse
the output properly with the json
Exporting and Importing
$ wf alert describe 1463413550083 -v 4 -f json | json severity SMOKE
What if we wanted to roll back to that alert? Of course, we could
set that single change back to the old value, but what
if we wanted to go back a number of revisions? Here’s how we’d do
$ wf alert describe 1463413550083 -v 4 -f json >alert-4.json $ wf alert delete 1463413550083 Soft deleting alert '1463413550083'. Deleted alert '1463413550083'. $ wf alert delete 1463413550083 Permanently deleting alert '1463413550083'. Deleted alert '1463413550083'. $ wf alert import alert-4.json Imported alert. created 1497275466684 minutes 2 name JPC Failed Services id 1497275466684 target firstname.lastname@example.org, status CHECKING inTrash false updateUserId email@example.com createUserId firstname.lastname@example.org lastProcessedMillis 1497275444832 pointsScannedAtLastQuery 0 createdEpochMillis 1497275466684 updatedEpochMillis 1497275466684 updaterId email@example.com creatorId firstname.lastname@example.org condition ts("dev.diamond.host.smf.svcs.maintenance", host="74a247a9-f67c-43ad-911f-fabafa9dc2f3") > 0 updated 1497275466684 severity SMOKE additionalInformation A service has failed. Log on to the box and see what it is deleted false
There’s the old alert, fully restored. It has a new
id, but that’s
okay. Everything significant is just the same. (You can also
import an alert, or any other object, over the top of an existing
one if you add the
Once an alert is exported you can, of course, do things to it before re-importing.
At my client’s site we have a user who has a number of environments: dev, staging, prod and so on. He created alerts for the first environment in the Wavefront console, then exported them and made them into ERB templates. Now, when he stands up a new environment, a script combines those templates with a few parameters to generate a whole new set of alerts, which it pushes to Wavefront. When he tears down an environment, a script deletes all alerts tagged with the environment being destroyed. Infrastucture as code, and alerts as part of your infrastructure.
If you don’t want to make templates (and I, personally, don’t), it’s probably simpler and cleaner to manipulate the original structured data. All you have to do is, in the language of your choice, load and parse some JSON, change what needs to be changed, dump it and re-import it.
To illustrate, here is a Ruby script which will read a JSON format alert from STDIN, change the condition, and dump the modified JSON to STDOUT.
#!/usr/bin/env ruby require 'json' alert = JSON.parse(STDIN.read) alert['condition'] = '0 > 1' puts alert.to_json
If I save that to an executable file called
modifier, I can run
$ wf alert describe -f json 1497275466684 | modifier | wf alert import -
and get a new alert with a new condition, leaving the old one in place. Obviously, it would be no more difficult to change any other aspect of the alert, or to source it from a version-controlled file rather than pulling it out of Wavefront.
To get all of your alerts as a single blob of JSON, run
$ wf alert dump -f json >all_alerts.json
If you wanted a subset of alerts, you could pipe
json. You can also use
-f yaml, should you
This bulk data can be re-imported with a standard
Bulk import/export can be very useful if you were migrating data
between clusters, but there are things to watch out for. For
instance, if you migrated your alert targets, the targets would all
get new IDs, which would mean a bulk import of alerts would likely
fail, as the old IDs don’t exist.
There’s no special syntax for a bulk
wf detects multiple
objects in the input, and deals with them automatically.
-M option lets you run other commands in a way which will produce
importable data. You could get a batch export of all your ‘JPC’ alerts
with a command like:
$ wf alert search name~JPC -f json -M >jpc_alerts.json
Exporting HCL for Use with Terraform
If you use Terraform, the Wavefront Terraform
can create your alerts, alert-targets and dashboards as part of a
stack. To make this easier, supplying
-f hcl to a
subcommand will export any of those objects in
HCL format, ready of pasting
straight into your Terraform configuration. (I discuss this further
What else can we do with the alerts CLI? Well, we can easily snooze
and unsnooze an alert. Let’s make an alert that’s always going to
fire. Save the following block of YAML as
--- name: test alert target: email@example.com, condition: "2 > 1" displayExpression: "" severity: SMOKE minutes: 2
Now import it. The CLI will happily import JSON or YAML, so long as the file has a sensible suffix.
$ wf alert import alert.yaml ... $ wf alert summary active 1 active_smoke 1 checking 9 snoozed 1 trash 15
Ooh, look, a firing alert! More info please!
$ wf alert firing 1497276280057 FIRING test alert
What do you know, it turns out that 2 is greater than 1. Good job we had an alert set up for that!
Snooze that alert for now.
$ wf alert snooze -T 10 1497276280057
Ten seconds later, and it turns out 2 is still greater than 1. Snooze it again, this time, indefinitely.
$ wf alert snooze 1497276280057
alert firing and
alert snoozed are deprecated now. Changes in
made it simple to add a more generic
currently sub-command. So
alert currently firing shows you all firing alerts, and
currently no_data will show you all the ones whose series have no
points over their last “would fire” interval. Valid alert states are
If you want to know what hosts are causing an alert to fire:
$ wf alert affected hosts 1594477854869 cube tornado
and if you just want to see who’s making noise anywhere in your estate, drop the alert ID, and you’ll see everything
$ wf alert affected hosts 1594477854869 ./wf alert affected hosts 1594309125807 cube 1594477854869 cube tornado
What Queries do I Have?
alert queries subcommand will show you the conditions used
across all your alerts. This can be useful if you’re thinking of
thinning out the metrics you collect.
$ wf alert queries 1459508340708 sum(deriv(ts(~collector.points.valid))) > 500 1464128764869 rate(ts("~agent.points.2878.sent", dc=home)) < 1 1476741941156 msum(3m, rate(ts("disk.error.*errors", !vendor=TSSTcorp))) 1489162558204 ts("zpool.*.cap") > 79
The final batch of
alert sub-commands are to do with tagging. It’s
probably easiest just to show you those:
$ wf alert tags 1497276280057 No tags set on alert '1497276280057'. $ wf alert tag add 1497276280057 example Tagged alert '1497276280057'. $ wf alert tag add 1497276280057 sysdef Tagged alert '1497276280057'. $ wf alert tags 1497276280057 example sysdef $ wf alert tag clear 1497276280057 Cleared tags on alert '1497276280057'. $ wf alert tags 1497276280057 No tags set on alert '1497276280057'. $ wf alert tag set 1497276280057 example sysdef numbers Set tags on alert '1497276280057'. $ wf alert tags 1497276280057 example numbers sysdef
Remember that most tags in Wavefront are one-dimensional: point tags
We’re finished for now, with our tour of the CLI alerting interface. All that remains is for us to not commit the cardinal sin of leaving an indefinitely snoozed alert.
$ wf alert unsnooze 1497276280057
I get a bit paranoid about having missed a firing alert, so I
sometimes run the CLI just to double-check I haven’t missed a
notification. To make my life marginally simpler I added a
$ wf alert firing 1459508340708 Point Rate 2018-02-17 01:47:39.929
Good job I did. I’m over my allocated point rate! I also like to be able to check that no one has been snoozing alerts instead of fixing them.
$ wf alert snoozed 1489162558204 Zpool usage 2017-11-17 11:03:12.922
Ooh, some cheeky so-and-so has used up all the disk space and silenced the alert so I didn’t find out!
The API calls alert targets “notificants”. The SDK echoes that, and the CLI follows the SDK. So, one manages one’s alert targets with the “notificant” command.
Alert targets are typically big things with lots of templating, so the CLI doesn’t provide a short-hand way of creating them in the way it does for, say, events, or derived metrics. You can still import them though. You just have to create a JSON or YAML description, likely starting from an existing one in the way we did with an alert earlier.
As well as
describe-ing alert targets, you can do all the usual
listing, deleting, updating, searching, and even testing.
$ wf notificant list CHTo475vsPzSaGhh WEBHOOK Slack alert webhook EKdKFv1rJ6ibahqI EMAIL alerts from lab machines T0i98AtVbs6Zkzlz EMAIL alerts from JPC production instances $ wf notificant test CHTo475vsPzSaGhh
You’ll have to trust me, but I promise that just popped up a Slack notification on my desktop.
A user can create up to twenty API tokens, and the
lets you manage your own tokens.
$ wf apitoken list fb83495d-9a44-26c5-fe41-1f7dd670734f $ wf apitoken create d8c5b877-d270-4990-9b5d-351015bf44c6 $ wf apitoken rename d8c5b877-d270-4990-9b5d-351015bf44c6 "example token" tokenID d8c5b877-d270-4990-9b5d-351015bf44c6 tokenName example token $ wf apitoken list fb83495d-9a44-26c5-fe41-1f7dd670734f d8c5b877-d270-4990-9b5d-351015bf44c6 example token $ wf apitoken delete d8c5b877-d270-4990-9b5d-351015bf44c6 Deleted api token 'd8c5b877-d270-4990-9b5d-351015bf44c6'.
Obviously you can’t create a token until you have a token, so the API isn’t quite up to full machine-generation of normal user accounts.
dashboard commands align with
alert ones. Obviously you
snooze a dashboard, but most of the others work just the
same. Dashboard descriptions can be h-u-g-e, so quite a lot of
information is dropped when you
describe one in human-readable
format. Everything I said about exporting and templating or
manipulating alerts applies just as well to dashboards.
There are some things you can do to dashboards that you can’t do to alerts. For instance, a user can have favourite dashboards. We have commands to manage these, carefully chosen to avoid trans-Atlantic spelling wars.
$ wf dashboard favs jpc-telegraf cube $ wf dashboard fav discogs discogs jpc-telegraf cube $ wf dashboard unfav discogs jpc-telegraf cube
(The SDK solves the spelling conundrum by aliasing
Dashboards now understand ACLs. These let you grant view or view-and-modify privileges to any users or user groups. (We’ll learn how to manage those in a while.)
By default, everyone can view and modify a dashboard. Let’s have a little play with privileges. You must specify accounts and groups by their IDs. It just happens that user account IDs are the same as their names. Let’s make a dashboard editable only by our two superstar 10xers, but viewable by everyone.
$ wf dashboard acls demo view and modify Everyone (2659191e-aad4-4302-a94e-9667e1517127) view <none> $ wf dashboard acl grant modify on demo to user firstname.lastname@example.org email@example.com view and modify Everyone (2659191e-aad4-4302-a94e-9667e1517127) firstname.lastname@example.org (email@example.com) firstname.lastname@example.org (email@example.com) view <none> $ wf dashboard acl revoke modify on demo from group 2659191e-aad4-4302-a94e-9667e1517127 view and modify firstname.lastname@example.org (email@example.com) firstname.lastname@example.org (email@example.com) view <none> $ wf dashboard acl grant view on demo to group 2659191e-aad4-4302-a94e-9667e1517127 view and modify firstname.lastname@example.org (email@example.com) firstname.lastname@example.org (email@example.com) view Everyone (2659191e-aad4-4302-a94e-9667e1517127)
Did you think the way I moved the
Everyone group was a bit
long-winded? Well, there’s some (I thought) unexpected behaviour
when you interact with the dashboard ACL API. Let’s continue the
above, and give the
Everyone group the right to modify that
$ wf dashboard acl grant modify on demo to group 2659191e-aad4-4302-a94e-9667e1517127 view and modify Everyone (2659191e-aad4-4302-a94e-9667e1517127) firstname.lastname@example.org (email@example.com) view <none>
You can see it’s removed the
view privilege. Fair enough,
is a member of
view and modify. But that means:
$ wf dashboard acl grant view on demo to group 2659191e-aad4-4302-a94e-9667e1517127 view and modify Everyone (2659191e-aad4-4302-a94e-9667e1517127) firstname.lastname@example.org (email@example.com) view <none>
You’ve asked for
Everyone to have
view, when the group already
had it, by virtue of having
view and modify. So simply adding
view doesn’t do anything. That’s why I removed it then added it.
If you get in a tangle,
wf will take you back to square one.
$ wf dashboard acl clear demo view and modify Everyone (2659191e-aad4-4302-a94e-9667e1517127) view <none>
You can’t do quite so many things with proxies.
search and so-on all work as for other commands, but
proxies can’t be tagged, or have ACLs or histories.
There are a couple of proxy-specific commands though. Proxies can be
renamed with – guess what –
proxy rename, and you can get a list
of proxy versions with…
$ wf proxy versions b75bf052-9985-407e-b90c-479e0134e261 4.35 Proxy on log.prod.wavefront-proxy 9997ac72-e755-4f8e-b3c6-1cdf2f991df8 4.35 Proxy on log.prod.wavefront-proxy 9662e8f1-2255-42cd-acf2-36f44170486f 4.35 Proxy on log.usprod.wavefront-proxy 8f93737a-5c21-45ea-8329-3adc4f80c215 4.35 Proxy on log.usprod.wavefront-proxy 88d181a0-2694-4886-b917-625a783e7783 4.35 Proxy on log.usprod.wavefront
Proxies are sorted with the most recent version at the top, descending. (The Go proxy doesn’t report a version, so any instances of that come right at the end.)
You probably won’t manipulate sources with the CLI, but if you want to, support is there. You can list, tag and describe them. That’s it.
The UI calls them “System Preferences”, but the API calls them
settings, and I tend to follow API conventions. The settings (or
system preferences) are defaults for new users. Things like the
group memberships or privileges a user has when they are invited to
I doubt anyone will find the
settings command particularly useful,
but it is here for completeness.
The CLI is able to interact more with events than with alerts, or
proxies, or dashboards, so we gain a couple of new subcommands in
$ wf event list $
What, no events? Well, no events in the last ten minutes, which is
the default view when you
list events. How about all events today?
$ wf event list -s 00:00 1497313265697:Alert Edited: No discogs update ENDED 1497310945968:Alert Snoozed: JVM Memory ENDED 1497310940168:Alert Deleted: test alert ENDED
Event names are, IMO a bit of a mess. They are the millisecond epoch
timestamp at which the event was created, joined, by a
:, to the
name of the event. When those names are pretty much free-form
strings like those above, it can get a little confusing. Let’s have
a look at that top one, remembering to quote the name.
$ wf event describe "1497313265697:Alert Edited: No discogs update" startTime 2017-06-13 00:21:05.697 endTime 2017-06-13 00:21:05.698 name Alert Edited: No discogs update annotations severity info type alert-updated userId firstname.lastname@example.org created 1495232095593 id 1497313265697:Alert Edited: No discogs update table sysdef updaterId System Event creatorId System Event canClose false creatorType SYSTEM canDelete false runningState ENDED
We can see that’s a system event. Something to know about system events is that you can’t delete them.
$ wf event delete "1497313265697:Alert Edited: No discogs update" API 400: Can only delete user events.
Let’s create an event. First, a couple of instantaneous events, occuring right this minute, because they’re the simplest kind.
$ wf event create -i BANG! ... $ wf event create -i BITE! -H shark ...
The first is a vague, floating-in-space event. It’s not attached to a
host, and to see it in your dashboards you’d have to turn on “Show
Events: All”. The second is attached to the host
shark, so it’ll
turn up on my Shark dashboard with no extra effort. You can attach
an event to as many hosts as you like.
Both those events could probably do with a bit more information, and
the CLI lets us specify severity (
-S) event type (
-T), and a
plain-text description of an event (
$ wf event create TORNADO! -H shark -S SEVERE -y unlikely \ -d "an unlikely event" Event state recorded at /var/tmp/wavefront/rob/1497366980092:TORNADO!. startTime 1497366980092 name TORNADO! annotations severity SEVERE type unlikely details an unlikely event id 1497366980092:TORNADO! table sysdef createdEpochMillis 1497366980746 updatedEpochMillis 1497366980746 updaterId email@example.com creatorId firstname.lastname@example.org createdAt 1497366980746 updatedAt 1497366980746 hosts shark isUserEvent true runningState ONGOING canDelete true canClose true creatorType USER
Notice that first line of output. The CLI has created, on the local
host, (not on
shark) a “state file”. This is a little memo of the
event ID, and every open event (i.e. one which is not instantaneous
and does not specify and end time) forces the creation of one. Those
state files work like a stack, and simply issuing an
command will pop the first one (that is, the last one that was
created) off the top of the stack, and close it. You can also supply
the name of an event to the
close command (just the name: no
timestamp part) and the last event opened with that name will be
closed. At any time you can see what events this host has open with
event show. Watch.
$ wf event show 1497366980092:TORNADO! $ wf event create test $ wf event create example $ wf event create example $ wf event create illustration $ wf event show 1497367580300:illustration 1497367359553:example 1497367333886:example 1497367298974:test 1497366980092:TORNADO! $ wf event close test $ wf event show 1497367580300:illustration 1497367359553:example 1497367333886:example 1497366980092:TORNADO! $ wf event close tornado No locally stored event matches 'tornado' $ wf event close 1497367333886:example $ wf event close TORNADO! $ wf event show 1497367580300:illustration 1497367359553:example $ wf event close $ wf event close $ wf event show No open events.
My most common use of
wf event is to wrap some command or other in
an event. I do this so often, I made a subcommand specifically for
$ wf event wrap -C 'stress --cpu 3 --timeout 1m' -T example "pointless busy work" Event state recorded at /var/tmp/wavefront/rob/1501109228938:pointless busy work. Command output follows, on STDERR: ---------------------------------------------------------------------------- stress: info:  dispatching hogs: 3 cpu, 0 io, 0 vm, 0 hdd stress: info:  successful run completed in 60s ---------------------------------------------------------------------------- Command exited 0 $ echo $? 0
Note that “output follows, on STDERR”.
wf takes all output,
standard out and standard error, from the wrapped command, and
dumps it to stderr. This is so, should you need to, you can separate
out the command output. In
event wrap mode,
wf exits whatever
the wrapped command exited. Here’s a chart showing the event.
(You have to hover over the event to see it.)
$ wf event describe "1501109228938:pointless busy work" id 1501109228938:pointless busy work name pointless busy work annotations type example details stress --cpu 3 --timeout 1m table sysdef startTime 2017-07-26 23:47:08.938 endTime 2017-07-26 23:48:10.492 createdAt 2017-07-26 23:47:09.721 createdEpochMillis 2017-07-26 23:47:09.721 updatedEpochMillis 2017-07-26 23:48:10.492 updaterId email@example.com creatorId firstname.lastname@example.org updatedAt 2017-07-26 23:48:10.492 isUserEvent true runningState ENDED canDelete true canClose true creatorType USER
You can see that
wf has put the command it wrapped into the
details field. If I had supplied an event description with
that would have been used instead.
I don’t use maintenance windows, as the systems I work on are built to tolerate the removal of pretty much any component. But, Wavefront does have good support for them, which the CLI covers. Creating a window is fairly simple:
$ wf window create -d 'demonstrating the CLI' -H shark 'example window'
You must supply a reason the window exists (with
-d) and a title
for the window, which is the final argument. You also have to give
Wavefront some way to connect a window to some sources. This can be
done with alert tags (using
-A), source tags (
-T), or host name
-H). These aren’t the CLI’s constraints, they’re the
Wavefront engine’s. So, the window above will stop any alerts firing
on any host whose name matches the string
shark. That’s nice for
me, because all the zones on that server have
shark as their
hostname prefix. (Yes,
shark is a pet: it lives in a cupboard in
my house.) You can mix and match tags and source names, and
AND them all together.
Note that I didn’t supply a start or end time for my window. Wavefront requires a start and end time when you create a window, and the CLI has filled them in for me, opening the window right now, and closing it in one hour.
$ wf window describe 1501844960880 id 1501844960880 reason demonstrating the CLI customerId sysdef createdEpochMillis 2017-08-04 12:09:20.880 updatedEpochMillis 2017-08-04 12:09:20.880 updaterId email@example.com creatorId firstname.lastname@example.org title example window startTimeInSeconds 2017-08-04 12:09:20 endTimeInSeconds 2017-08-04 13:09:20 relevantHostNames shark eventName Maintenance Window: example window runningState ONGOING
If I wish, I can extend it. Let’s give ourselves another hour.
$ wf window extend by 1h 1501844960880 $ wf window describe 1501844960880 | grep endTime endTimeInSeconds 2017-08-04 14:09:20
Or we can close it bang on 2pm
$ wf window extend to 14:00 1501844960880 $ wf window describe 1501844960880 | grep endTime endTimeInSeconds 2017-08-04 14:00:00
Or just close it immediately.
$ wf window close 1501844960880 id 1501844960880 reason demonstrating the CLI customerId sysdef createdEpochMillis 1501844960880 updatedEpochMillis 1501845460225 updaterId email@example.com creatorId firstname.lastname@example.org eventName Maintenance Window: example window title example window startTimeInSeconds 1501844960 endTimeInSeconds 1501845458 relevantHostNames shark runningState ENDED
To see which windows are ongoing, use
wf window ongoing, and to
see which are coming up soon, use
wf window pending. By default,
pending shows windows which will open in the next 24 hours, but it
takes an optional “hours” argument. So, what windows are coming up
in the next two days?
$ wf window pending 48 No maintenance windows in the next 48.0 hours.
Like I said, I don’t use them,
You can import and export maintenance window objects, just like everything else.
You can manage derived
metrics with the
derivedmetrics command. As well as all the usual deleting,
describing, importing, and whatnot, you can create derived metrics
on the command line. You have to supply a name for the derived
metric, along with the actual metric. So something like:
$ wf derivedmetric create my_metric 'aliasMetric(ts(real.series), "alias")' $ wf derivedmetric list 1529944840652 my_metric
Like most things in Wavefront, derived metrics can be tagged, and the
derivedmetric create subcommand lets you do this on the fly, as
well as specifying a description; adjusting the interval at
which the metric runs on the cluster; specifying the amount of
time over which the metric is created; and whether or not to include
obsolete metrics in the calculations.
Derived metrics support history and soft-deleting.
Cloud Integrations, Webhooks, External Links and Saved Searches
savedsearch commands are
simpler than those we’ve seen so far, becaue the API doesn’t allow
tagging or soft-deleting of those resource types. The CLI still
lets you list, describe, delete and import them though, and each has
properties you can
search subcommand works for all
of them too.
Now let’s look at the “oddball” commands.
source command lets you manage tags and descriptive strings
for any of your sources. In Wavefront “sources” usually equate to
hosts or containers, but they don’t have to. I have some
applications which identify as a source, because I don’t care where
they run, only what they say.
You will have lots of sources, so the output of
sources list will
likely be paginated. (If it is, it says so.) You can use
-L to set the offset (starting point) and limit of the page you
view. These options work for all
list sub-commands, but when
-o should refer to a source name. For everything
else it is a numerical offset. This reflects the way the API works.
source list does not show sources which are “hidden”
(which usually means they are very old) or Wavefront’s own sources
(the things that look like
If you want to see these, supply the
The hidden and internal sources are filtered out by the CLI
after the API call is made, but the
-L flags are set
before the call. This can mean it can get a confusing working
through the pages. You’re probably better off using
search, or using
-a and dealing with it.
$ wf source describe shark-ws id shark-ws sourceName shark-ws hidden false description workstation zone tags ~status.errortrue zone true solaris true $ wf source clear shark-ws status result OK code 200 $ wf source describe shark-ws id shark-ws sourceName shark-ws hidden false $ wf source description set shark-ws "workstation zone" status result OK code 200 $ wf source tag add shark-ws solaris Tagged source 'shark-ws'. $ wf source tag add shark-ws zone Tagged source 'shark-ws'. $ wf source describe shark-ws id shark-ws sourceName shark-ws hidden false description workstation zone tags solaris true zone true
You won’t use this one very much, but it’s in the API, so the CLI covers it.
wf message list shows any messages you may have. You get these, for
instance, if your cluster is going to be upgraded, and you see them
across the top of the page when you log in to the UI.
You can read a message with
wf message read <id>, or by using
wf message list -l, which will also show you things like the scope
and severity of the message.
If you use
read, the message will be marked as “read”, and not
show up when you list messages. You can do this manually with the
Read messages can still be shown with
list -a, and
They appear to age out after
endEpochMillis has passed.
$ wf message list CLUSTER::743cvsHu Wavefront Upgrade Notification $ wf message read CLUSTER::743cvsHu Wavefront Upgrade Notification ------------------------------ Wavefront is upgrading to the latest version within the next two (2) weeks. -Wavefront Customer Success email@example.com $ wf message list $ wf message list -a CLUSTER::LmPJTdQ8 Wavefront Upgrade Notification
Though accounts can have permissions directly assigned, roles are the preferred way to manage privileges.
$ wf role list 07fc5cdd-0979-489e-8f70-325f39d15e55 admin (auto-created) 0 accounts 1 groups 0a42adf6-e738-4c5d-9e53-fd10bd979a31 testrole 1 accounts 0 groups 6009b7e7-aa04-43bb-8c71-caceda6dfac9 proxy_group (auto-created) 0 accounts 1 groups 6180fbe8-8dbb-4e9c-b73d-8c53424771ef example role 1 accounts 2 groups
Hopefully you, the user, aren’t too offended by
1 groups. It looks a mess
s is dynamic, and
group(s) is too fussy.
We can, of course
describe any of those roles and see their various
properies. But those properties don’t necessarily include all the groups or
accounts which have the given role. But
wf can help.
$ wf role accounts 6180fbe8-8dbb-4e9c-b73d-8c53424771ef sa::explorer $ wf role groups 6180fbe8-8dbb-4e9c-b73d-8c53424771ef 2352f992-1a52-42c3-9206-9ef7c838a5a0 explorer 64fc3264-fc43-4b7e-87fc-139c9ed29c2a demo_group
It can tell you a role’s permissions too:
$ wf role permissions 6180fbe8-8dbb-4e9c-b73d-8c53424771ef Role '6180fbe8-8dbb-4e9c-b73d-8c53424771ef' has no permissions.
That’s not a lot of use is it? Let’s add some permissions. But what the heck are they?
Remember I said the
settings command wasn’t much use? It does serve at least
one useful purpose.
$ wf settings list permissions agent_management alerts_management application_management batch_query_priority dashboard_management derived_metrics_management embedded_charts events_management external_links_management host_tag_management ingestion metrics_management user_management
If any seem vague:
$ wf settings list permissions --long groupName alerts_management displayName Alerts description Users with this permission can manage alerts, maintenance windows, and alert targets. requiredDefault false -------------------------------------------------------------- groupName batch_query_priority displayName Batch Query Priority description Users with this permission will run at a lower priority level for queries (mainly for users for role accounts intended for reporting purposes) requiredDefault false ...
$ wf role grant alerts_management to 6180fbe8-8dbb-4e9c-b73d-8c53424771ef Granted 'alerts_management' permission to '6180fbe8-8dbb-4e9c-b73d-8c53424771ef'. $ wf role grant dashboard_management to 6180fbe8-8dbb-4e9c-b73d-8c53424771ef Granted 'dashboard_management' permission to '6180fbe8-8dbb-4e9c-b73d-8c53424771ef'. $ wf role permissions 6180fbe8-8dbb-4e9c-b73d-8c53424771ef alerts_management dashboard_management
Nice. We can take permissions away too.
$ wf role revoke dashboard_management from 6180fbe8-8dbb-4e9c-b73d-8c53424771ef Revoked 'dashboard_management' permission from '6180fbe8-8dbb-4e9c-b73d-8c53424771ef'. $ wf role permissions 6180fbe8-8dbb-4e9c-b73d-8c53424771ef alerts_management
Let’s see who our role is assigned to.
$ wf role accounts 6180fbe8-8dbb-4e9c-b73d-8c53424771ef sa::explorer $ wf role groups 6180fbe8-8dbb-4e9c-b73d-8c53424771ef 2352f992-1a52-42c3-9206-9ef7c838a5a0 explore 64fc3264-fc43-4b7e-87fc-139c9ed29c2a another_test
I don’t know who the heck any of those are. Let’s get rid of them. You can treat groups and accounts just the same, and specify many at once.
$ wf role take 6180fbe8-8dbb-4e9c-b73d-8c53424771ef from \ sa::explorer \ 2352f992-1a52-42c3-9206-9ef7c838a5a0 \ 64fc3264-fc43-4b7e-87fc-139c9ed29c2a Took '6180fbe8-8dbb-4e9c-b73d-8c53424771ef' from 'sa::explorer', '2352f992-1a52-42c3-9206-9ef7c838a5a0', '64fc3264-fc43-4b7e-87fc-139c9ed29c'.
You can do all the usual
setting of properties,
importing of roles. It’s also easy to make a new role, with whatever
permissions you wish it to have.
$ wf role create -d "CLI example" -p ingestion -p embedded_charts example_role id 685b3d44-1c82-46c3-bfde-ea3223512bd0 name example_role sampleLinkedGroups <none> linkedGroupsCount 0 sampleLinkedAccounts <none> linkedAccountsCount 0 permissions embedded_charts ingestion customer sysdef description CLI example lastUpdatedMs 1594126095896 lastUpdatedAccountId firstname.lastname@example.org
We can now assign that group directly to accounts, or, as is probably better, to groups. Let’s have a look at managing those.
Let’s start by seeing what groups I already have defined.
$ wf usergroup list 2659191e-aad4-a34d-a94e-9667e1517127 Everyone 4
The long UUID string on the left is the group ID, which you use in almost all
wf usergroup commands.
Every cluster has the
Everyone group, but your
have a different ID to mine. By default,
Everyone has no
The other columns are the name of the group, and the number of users in it. Who are those users? (None of these are real, because I don’t want all my users e-mail addresses being harvested for spam.)
$ wf usergroup users 2659191e-aad4-a34d-a94e-9667e1517127 email@example.com firstname.lastname@example.org email@example.com firstname.lastname@example.org
In the old days you could assign permissions to a usergroup. That’s not possible any more: you assign permissions to roles, and roles to usergroups.
You can specify the roles a group has when you create it, but you can’t currently specify the users in the group. This is a limitation of the API.
$ wf usergroup create -r 685b3d44-1c82-46c3-bfde-ea3223512bd0 "normal users" id f8dc0c14-91a0-4ca9-8a2a-7d47f4db4672 userCount 0 permissions alerts_management dashboard_management events_management customer sysdef createdEpochMillis 1550683825337 name normal users
Now we have a group, we can put some of our users into it. We can add as many users and/or service accounts as we like.
$ wf usergroup add to b4ea2152-a417-430e-af43-d1b7fa69dfbb email@example.com sa::explorer Added 'firstname.lastname@example.org', 'sa::explorer' to 'b4ea2152-a417-430e-af43-d1b7fa69dfbb'. $ wf usergroup add role b4ea2152-a417-430e-af43-d1b7fa69dfbb 0a42adf6-e738-4c5d-9e53-fd10bd979a31 Added '0a42adf6-e738-4c5d-9e53-fd10bd979a31' to 'b4ea2152-a417-430e-af43-d1b7fa69dfbb'. $ wf usergroup users b4ea2152-a417-430e-af43-d1b7fa69dfbb email@example.com sa::explorer $ wf usergroup roles b4ea2152-a417-430e-af43-d1b7fa69dfbb 0a42adf6-e738-4c5d-9e53-fd10bd979a31 $ wf usergroup permissions b4ea2152-a417-430e-af43-d1b7fa69dfbb ingestion
You could complain about the ordering of the arguments here.
The commands would read better you could, for instance, write
add role role-id to group-id.
But I couldn’t make
docopt, which parses
wf’s command lines, work
in that way. I tried making the users be defined with a repeated option, but
that didn’t seem right: users are not an option in a command called
add. User interface design is full of compromises, and this is one. I hope
you don’t find it too objectionable.
You’ll find that re-running a lot of the commands produces the same output. The API guarantees idempotency with a declarative approach. You don’t so much request the removal of a user from a group as assert that said user is not in the group. Therefore you shouldn’t care whether the user was actually removed or never existed, only that it isn’t there now.
You can, of course,
delete groups, export them with
and re-import them with
import. After an import, the group gets a
new ID, and will not have any users assigned to it. (Group
membership is an attribute of a user, rather than users being an
attribute of a group.)
In version 7 of
wf account replaces
wf user. This
reflects Wavefront’s deprecation of the
user API. I was glad to see it go:
it didn’t work in quite the same way as other paths, and the CLI and SDK had
to do some messy behind-the-scenes work to make user management function
similarly to everything else.
We’ve seen the
commands enough by now to now what they do.
account has a lot of other
things we’ve kind of seen too:
permissions shows you the per
Some of the account commands include the word
user. This is to help
differentiate between accounts for humans and accounts for machines, which
we’ll come to later.
As of version 7, when you
invite a user, that user has no
permissions unless you explicitly define them. The API’s default behaviour
is, if given no permissions, is to add some. I chose to override this.
revoke subcommands are at your disposal, I would not
recommend using permissions at all. Use roles, via usergroups.
invite let you fully automate user creation,
for instance in the creation of machine accounts. It is not
currently possible to validate an account or create an API token via
Wavefront now supports service accounts. These are great for proxies, and for
getting your tooling properly wired up to Wavefront. There’s almost, but not
quite, full API coverage for service accounts, and this is reflected in the
CLI. Let’s make a service account for our proxies, granting it the
$ wf serviceaccount list You have no service accounts. $ wf serviceaccount create -p ingestion sa::proxy identifier sa::proxy tokens <none> userGroups id a7d2e651-cec1-4154-a5e8-1946f57ef5b3 name Everyone permissions <none> customer sysdef properties nameEditable false permissionsEditable true usersEditable false description System group which contains all users active true groups ingestion $ wf serviceaccount apitoken list sa::proxy Account does not have any API tokens. $ wf serviceaccount apitoken create -N "proxy ingestion token" sa::proxy tokenID 416f0fc5-628f-4b2d-b3c8-34a9aa6cf74d tokenName proxy ingestion token $ wf serviceaccount describe sa::proxy identifier sa::proxy tokens tokenID 416f0fc5-628f-4b2d-b3c8-34a9aa6cf74d tokenName proxy ingestion token userGroups id a7d2e651-cec1-4154-a5e8-1946f57ef5b3 name Everyone permissions <none> customer sysdef properties nameEditable false permissionsEditable true usersEditable false description System group which contains all users active true groups ingestion
Assigning permissions to users is are fine, but I think user groups are better. So let’s make a proxy group and move our service account into that, revoking the permission.
$ wf usergroup create -p ingestion proxy_group id afa04fcd-5e27-495b-9ebc-c732aba42438 name proxy_group users <none> userCount 0 permissions ingestion customer sysdef createdEpochMillis 1569843554718 $ wf serviceaccount join sa::proxy afa04fcd-5e27-495b-9ebc-c732aba42438 a7d2e651-cec1-4154-a5e8-1946f57ef5b3 (Everyone) afa04fcd-5e27-495b-9ebc-c732aba42438 (proxy_group) $ wf serviceaccount revoke ingestion from sa::proxy Revoked 'ingestion' from 'sa::proxy'.
Now I’ve shown you, let’s clean up.
$ wf serviceaccount apitoken delete sa::proxy 416f0fc5-628f-4b2d-b3c8-34a9aa6cf74d Deleted API token '416f0fc5-628f-4b2d-b3c8-34a9aa6cf74d'.
We can delete the service account too.
$ wf serviceaccount delete sa::proxy Deleted service account 'sa::proxy'.
Ingestion Policies and Usage
Wavefront now has a feature to show you who is responsible for what proportion of your ingested point rate.
This is done by assigning accounts an ingestion policy. Any system account or normal user can belong to one or none ingestion policies. Wavefront provides you with dashboards to see how much of your point rate is attributable to each policy.
Creating an ingestion policy is very simple. Give it a description with
you’ll be glad you did one day.
$ wf ingestionpolicy create -d "example ingestion policy" example-policy id example-policy-1579802191862 name example-policy sampledUserAccounts <none> userAccountCount 0 sampledServiceAccounts <none> serviceAccountCount 0 customer sysdef description example ingestion policy lastUpdatedMs 1579802191878 lastUpdatedAccountId firstname.lastname@example.org
We can add any number of users or system accounts.
$ wf user list email@example.com firstname.lastname@example.org $ wf serviceaccount list sa::proxy $ wf ingestionpolicy add user another-ingestion-policy-1579538401492 \ email@example.com sa::proxy
Though you can, of course,
describe an ingestion policy,
wf gives you a
convenience command for showing which users belong to a given policy, and an
inverse operation to show which policy a given user comes under.
$ wf ingestionpolicy members example-policy-1579802191862 firstname.lastname@example.org sa::proxy $ wf ingestionpolicy for sa::proxy example-policy-1579802191862
You can get a CSV output of usage breakdown with
$ wf usage export csv
-f json won’t work with this: it’s a limitation of the Wavefront
metric lets you find out when a metric was last reported. The
output is sorted on the time, with the most recent first.
$ wf metric describe wavefront-proxy.host.uptime.uptime i-0b10ff25afd0e0c7d 2017-06-13 21:34:38.000 i-0c568ca14f72738a6 2017-06-13 20:56:03.000 i-05bc5822132c5863c 2017-06-13 18:58:15.000 i-059184d32a443b326 2017-06-13 13:42:37.000 i-014e5eb7991d97d4e 2017-06-11 03:14:21.000 i-0c425b83f5430dd13 2017-06-10 18:05:00.000 i-0fc90132760807425 2017-06-09 10:47:14.000 i-01bfb02a7c3ad843e 2017-06-07 23:35:28.000 i-0b2fa0060fc8eae88 2017-06-07 23:31:34.000 i-05f8817d3ac61bfab 2017-06-07 17:27:49.000
You can pattern-match your request with the
$ wf metric describe wavefront-proxy.host.uptime.uptime -g "i-05*" i-05bc5822132c5863c 2017-06-13 18:58:15.000 i-059184d32a443b326 2017-06-13 13:42:37.000 i-05f8817d3ac61bfab 2017-06-07 17:27:49.000
/metric API seems a little brittle at the moment, and throws a
500 if you search for a metric which does not exist. The CLI
dutifully reports this error.
wf also exploits an undocumented API endpoint to offer something akin to the
UI’s metric browser. I can, for instance, find out what metrics I have
$ wf metric list under dev dev.test.a dev.test.b
You can even ask for a list of all the metrics your cluster knows about:
$ wf metric list all
But, I really wouldn’t recommend you do that. If you imagine your metrics as
wf must make an API call for every single node of that tree. This
metric list commands can take a very, very long time to complete.
I’d like to see an official API path to browse metrics, with recursion done on the server. If you agree, lobby your Wavefront representetive!
wf also speaks to another unofficial API:
spy. This, as I understand it,
connects you to a single node of your Wavefront cluster, and shows a sampling
of the data flowing into it. You can spy on data points, histograms, traces,
or new source IDs. For instance, (and with lines folded for formatting)
$ wf spy opi "~proxy.push.2003.duration.rate.m1" source="log.prod.wavefront-proxy" 1581977100000 309.100850087495 "processId"="2204f9eb" "hcam.prod.yapp.app.gauges.buffers.direct.count" source="i-050ee5e3584853963" 1581977145000 98.0 "accountId"="308487525487" "product"="hcam" "environment"="prod" "role"="yapp" "_wavefront_source"="proxy::hcam.prod.i-0cc9b726ab232b03e" "hcam.prod.yapp.app.gauges.memory.pools.Compressed-Class-Space.committed" source="i-050ee5e3584853963" 1581977145000
wf shows you the unexpurgated data it gets from the API, but if
you want to use the data in some form of investigation, it may be useful to
--timestamp) option, which will drop a local timestamp
into the output ahead of each chunk of data. You can adjust the sampling rate
too, but it tops out at 5% of the data flowing to the node.
If you need detailed spy information, I’d recommend the far more sophisticated
query command has quite a lot of options (common ones removed for
wf query aliases [-DV] [-c file] [-P profile] wf query [-DnVm] [-c file] [-P profile] [-E endpoint] [-t token] [-g granularity] [-s time] [-e time] [-f format] [-WikvO] [-S mode] [-N name] [-p points] [-F options] <query> wf query raw [-DnVm] [-c file] [-P profile] [-E endpoint] [-t token] [-H host] [-s time] [-e time] [-f format] [-F options] <metric> wf query run [-DnVm] [-c file] [-P profile] [-E endpoint] [-t token] [-g granularity] [-s time] [-e time] [-f format] [-F options] [-WkivO] [-S mode] [-N name] [-p points] <alias> Options: -E, --endpoint=URI cluster endpoint -t, --token=TOKEN Wavefront authentication token -g, --granularity=STRING query granularity (d, h, m, or s) -s, --start=TIME start of query window -e, --end=TIME end of query window -N, --name=STRING name identifying query -p, --points=INTEGER maximum number of points to return -i, --inclusive include matching series with no points inside the query window -v, --events include events for matching series -S, --summarize=STRING summarization strategy for bucketing points (mean, median, min, max, sum, count, last, first) -O, --obsolete include metrics unreported for > 4 weeks -H, --host=STRING host or source to query on -F, --format-opts=STRING comma-separated options to pass to output formatter -k, --nospark do not show sparkline -W, --nowarn do not show API warning messages
Let’s run a simple query.
$ wf query 'deriv(ts("nfs.server.v4.read"))' name deriv(ts("nfs.server.v4.read")) query deriv(ts("nfs.server.v4.read")) timeseries label nfs.server.v4.read sparkline > ▁ ▂▁ ▃ ▂▁ ▃ ▁ ▂ < host shark tags env lab data 2018-06-25 17:10:29 0.0 17:10:39 0.0 17:10:49 0.0 17:10:59 0.0 17:11:09 0.0 17:11:19 0.0 17:11:29 0.0 17:11:39 0.5 17:11:49 0.0 17:15:59 0.0 17:16:09 0.0 17:16:19 0.0 17:16:29 0.6 17:16:39 0.4 17:16:49 0.0 17:16:59 0.8 17:17:09 0.0 ...
granularity is an important option. It lets you select the
bucket size Wavefront will use to aggregate data. If you don’t
supply a granularity, the CLI will try to work out the right one
based on the size of the time window you give. And if you don’t give
a time window, it will use the last ten minutes.
The Wavefront API expects the query window to be defined by start
and end times in epoch milliseconds, but the CLI will try to convert
any time format you give it, using Ruby’s
Times as loosely defined as
Saturday may well work, but
sometimes Ruby will assume
Saturday means the next one, not the
last one, so choose wisely!
sparkline is a bit of a novelty. It uses Unicode blocks, which
severely limits its range. If it annoys you,
-k turns it off.
As well as specifying the granularity of the point buckets (just
like the UI does, dependent on its canvas size), you can select the
strategy used on the values in those buckets. Like the UI, the
default strategy is
MEAN, but the
-S option lets you specify
LAST or any of the others offered by the UI.
raw sub-command requires a host and a metric path - not a
time-series expression. It gives you the raw values for that metric,
on that host, over a given range.
$ wf query raw 'lab.dev.host.nfs.server.v4.read' -H shark -s 13:00 -e 13:01 2017-06-14 12:00:06.000 127493.0 12:00:16.000 127493.0 12:00:26.000 127493.0 12:00:36.000 127493.0 12:00:46.000 127493.0 12:00:56.000 127493.0
Start and end times don’t have to be absolute. As of version 2.1.0
of the CLI, you can specify relative times. So you can run a query
over a window from “two hours ago” to “one hour ago”, with
-e -1h. Valid time units are
I hope are self-explanatory. Because these are relative ranges, the
CLI makes no attempt to compensate for any daylight saving or
You can specify future times as
+2.5h or similar. This is useful
for maintenance windows, but if you try to see into the future on a
query, the Wavefront API will, not unreasonably, throw an exception.
Storing Queries with Aliases
Imagine if you got a bit obsessive over running that NFS query.
You’d soon get tired of typing it in, and remembering to balance the
brackets and the quotes. Handily,
wf lets you “alias” commonly
used queries. To set up an alias called
nfs for the above query,
you would add this to the relevant stanza of your
q_nfs = deriv(ts("lab.dev.host.nfs.server.v4.read"))
Then to run the query (with default granularity and time windowing) you’d just do
$ wf query run nfs ...
You can, of course, specify all the normal query options with an
alias. The syntax, as I’m sure you noticed, is that the alias name
you’d use must be prefixed with
q_. This is a workaround for the
limitations of INI files, which don’t let you nest sections. (At
least, not in Ruby’s understanding of them.)
To see what aliases you have configured, you can just run
$ wf query aliases nfs
Query Output Formats
wf query can present its results in all the formats the other
commands use, but it supports a number of additional output formats.
Native Wavefront Output
wavefront writes out the points in native Wavefront wire format.
It works for timeseries and raw queries.
$ wf query 'ts("solaris.network.obytes64", environment=production)' -f wavefront solaris.network.obytes64 121037323749.0 1533754102 source=wf-blue env="prod" nic="net0" solaris.network.obytes64 121037670562.0 1533754122 source=wf-blue env="prod" nic="net0" solaris.network.obytes64 121038023454.0 1533754142 source=wf-blue env="prod" nic="net0" ... $ wf query raw -H www-blue 'solaris.network.obytes64' -f wavefront solaris.network.obytes64 1219563430.0 1533751241000 source=www-blue nic="net0" role="sinatra" solaris.network.obytes64 1219563982.0 1533751261000 source=www-blue nic="net0" role="sinatra" ...
You might have noticed that the timeseries query reports the
point timestamp as epoch seconds, whereas a
raw query phrases them
as epoch milliseconds. Don’t worry about it: the proxy accepts both.
You can pipe this data straight back into a proxy using
wf itself. This is great for cluster migrations. Or, if you
modify the data in-flight, perhaps with
awk, you can copy
data to new metric paths, or amend tags.
$ wf query 'ts("solaris.network.obytes64")' -f wavefront \ | sed 's/solaris/smartos/' \ | nc wf-proxy 2878
csv format outputs points as a CSV table. By default no column
headers are printed; values are not quoted unless they contain
whitespace, a comma or a soft quote; and point tags have their
values printed but not their keys.
By using the
-F option you can change all these things.
a comma-separated list of keywords:
headers will print the CSV
quote will soft-quote every value; and
print point tags as
$ wf query 'ts("solaris.network.obytes64")' -f csv -F headers | sed 2q path,value,timestamp,source,platform,nic,dc solaris.network.obytes64,192820246.33333334,1561651740,www-green,JPC,net0,eu-ams-1 $ wf query 'ts("solaris.network.obytes64")' -f csv -F tagkeys,quote | sed 1q "solaris.network.obytes64","192852172.33333334","1561651860","www-green","platform=JPC","nic=net0","dc=eu-ams-1"
Should you forget, running
wf query --help will tell you all of
write command has a lot of options and capabilities, so I
wrote a separate article about sending metrics from the
Code and Contributions
The CLI, and the SDK on which is built, are open source, available under a BSD license.
Contributions, bug reports, and feature requests are always welcome.