- Notifications
You must be signed in to change notification settings - Fork17
A library to easily send request events and traces tohttps://honeycomb.io
License
cloudflare/workers-honeycomb-logger
Folders and files
| Name | Name | Last commit message | Last commit date | |
|---|---|---|---|---|
Repository files navigation
Honeycomb logger is a simple library that lets you extremely easily export runtime information from yourCloudflare Workers intoHoneycomb.Honeycomb is an observability platform which allows you to query and graph across any (number of) dimension(s) you have in your data. So you can for example graph request duration for 200 response codes, for GET requests, to a particular URL, for a particular customer.
Or you can drill into an entire trace of a request that errored out, including all subrequests.
Table of Contents
- Prerequisites
- Getting Started
- Config
- Adding logs and other data
- Traces
- Dynamic Sampling
- Cloudflare Pages Support
This script does not yet work for scheduled workers. It supports Workers in both theaddEventListener and module syntaxes, including Durable Objects. For more information on Durable Object support, see below.
Installation is done via the usualnpm install @cloudflare/workers-honeycomb-logger oryarn add @cloudflare/workers-honeycomb-logger.
The next two things you need are a Honeycomb API key and a dataset name. You can pick any dataset name you like, Honeycomb will automatically create a new dataset if it sees a new name.
The next step depends on if you are using theaddEventListener format or the new module format (required for Durable Objects)
Automatically unpack nested JSON setting in the dataset settings inhttps://ui.honeycomb.io to be able to query over all the data that is automatically included.
To configure the package, you need to wrap your listener with the honeycomb logger. So if your current code looks something like this:
addEventListener('fetch',(event)=>{event.respondWith(handleRequest(event.request))})functionhandleRequest(request){//your worker code.}
You can change that to:
import{hc}from'@cloudflare/workers-honeycomb-logger'consthc_config={apiKey:'<api_key>',dataset:'<my-first-dataset>',}constlistener=hc(hc_config,(event)=>{event.respondWith(handleRequest(event.request))})addEventListener('fetch',listener)functionhandleRequest(request){//your worker code.}
You are now good to go. Read through the config section to see what else you can configure.If you run into any problems, here is afull example of an addEventListener configured worker
If you are using the module syntax, the setup is slightly different. But you still have to wrap the worker.
If your code was something like this:
exportdefault{asyncfetch(request,env,ctx){returnnewResponse('Hello world!',{status:200})},}
You would change that to something like:
constworker={asyncfetch(request,env,ctx){returnnewResponse('Hello world!',{status:200})},}constconfig={apiKey:'__HONEYCOMB_API_KEY__',// can also be provided by setting env var HONEYCOMB_API_KEYdataset:'my-first-dataset',// can also be provided by setting env var HONEYCOMB_DATASET}exportdefaultwrapModule(config,worker)
You are now good to go. Read through the config section to see what else you can configure.If you run into any problems, here is afull example of a module syntax configured worker
If you don't want to hardcode the API key manually, you can also set the environment variablesHONEYCOMB_API_KEY andHONEYCOMB_DATASET and they will override whatever is inserted intowrapModule.
If you are using Durable Objects, we support automatic distributed tracing across both your workers and your Durable Objects.You need to wrap your Worker code with themodule system configuration above, and then wrap your durable object as well.
You can do it like this:
import{Counterascounter_do}from'./counter.mjs'constworker={asyncfetch(request,env){// handleRequest},}constconfig={apiKey:'__HONEYCOMB_API_KEY__',// can also be provided by setting env var HONEYCOMB_API_KEYdataset:'my-first-dataset',// can also be provided by setting env var HONEYCOMB_DATASET}constCounter=wrapDurableObject(config,counter_do)exportdefaultwrapModule(config,worker)export{Counter}
If you run into any problems, here is afull example of a durable object configured worker
The config object can take a few extra parameters to add more detail to the events that are sent to Honeycomb or configure other aspects.
interfaceConfig{apiKey?:string//The honeycomb API Key, can be specified as env var HONEYCOMB_API_KEYdataset?:string//The name of the dataset, can be specified as env var HONEYCOMB_DATASETacceptTraceContext?:boolean//Do you want to accept automatic TraceContext information from clients? Defaults to 'false'data?:any//Any data you want to add to every request. Things like service name, version info etc.redactRequestHeaders?:string[]//Array of headers to redact. Will replace value with `REDACTED`. default is ['authorization', 'cookie', 'referer'].redactResponseHeaders?:string[]//Array of headers to redact. Will replace value with `REDACTED`. default is ['set-cookie'].sampleRates?:SampleRates|SampleRateFn//Either an object or function that configured sampling ([See below](#dynamic-sampling))sendTraceContext?:boolean|RegExp//set this to true to send a TraceContext with all fetch requests. With a Regex, we will check the URL against the regex first. Defaults to 'false'serviceName?:string//The serviceName you want to see in Honeycomb. Defaults to 'worker'debugLog?:boolean// console.log response info for Honeycomb requests. Defaults to false}
NOTE: In previous versions there were methods for parsing request and responses, but this becomes an issue when reading the body of the request or the response, so it has been removed. If you want add any information you can do so with thetracer.addData method described below.
If you want to add any other data or logs to the current request, you can use thetracer.addData(data: object) andtracer.log(message: string) methods.You can get a reference to the tracer either on the request object, or the second argument in the listener.
asyncfunctionhandleRequest(request:TracerRequest){request.tracer.log('handling request')constcustomer=parseCustomer(request)request.tracer.addData({customer:customer.name})return ...}
Honeycomb has a concept of a trace, which is a hierarchial representatation of multiple, related events. The Cloudflare Worker Honeycomb logger supports trace events for subrequests (outgoing HTTP fetch requests in your worker) like this:
To be able to associate the a subrequest with the correct incoming request, you will have to use the fetch defined on the tracer described above. The method on the tracer delegates all arguments to the regular fetch method, so thetracer.fetch function is a drop-in replacement for allfetch function calls.
Example:
asyncfunctionhandleRequest(request:Request){returnrequest.tracer.fetch('https://docs.honeycomb.io/api/events/')}
If you want to enable automatic distributed traces from your client to your backend, you can enable theacceptTrace property to participate in the trace started by your client andsendTraceContext to automatically send through theTraceContext to your backend.This only works on thefetch methods that are defined on thetracer object and is compatible with any of the official Honeycomb beeline frameworks.Remember that you do have to use the samedataset for all systems to be able to assemble your trace in Honeycomb.
One of the challenges with storing all this information per request is that when you scale that up to past tens of millions of requests as month, it becomes more and more expensive. But at the same time you are almost certainly not very interested in the vast majority of the events. Which is why Honeycomb supports sampling. Sending only a portion of the events there. The problem with doing simple sampling (like sending only 1 in 10 requests for example), is that you lose a lot of events that happen rarely. So Honeycomb and this library support dynamic sampling.The easiest sampling that you can configure is by response code. So you can configure to keep only 1 in 10 responses code 200s, and keep all 5xx.
The number is technically how many requests it is representative for. An easier way to remember is that a request has a 1/x chance to be sampled.
exportinterfaceSampleRates{'1xx':number'2xx':number'3xx':number'4xx':number'5xx':numberexception:number}
In an example:
consthc_config={api_key:'abcd',dataset:'my-first-dataset',sampleRates:{'1xx':20,'2xx':10,'3xx':5,'4xx':2,'5xx':1,exception:1,},}
This configures the library to only send 1 in 10 requests with a response code in the 200s, but keep all errors; both 500s and exceptions.
If you want more fine grained control over what is sampled, you can supply a function that takes the entiredata object of the outer event.This is the exact object that will be sent to Honeycomb, so you can see what values are available there.This includes all information that you have added withrequest.tracer.addData()
Please note that theHeaders for both request and response objects are converted into aRecord<string, string>. So if you want to check thecontent-type headers in the response, you can dodata.response.headers['content-type'].
Examples
functionsampleRates(data:any):number{if(data.request.url.contains('/healthcheck')){return0}elseif(data.customer.plan==='enterprise'){return5}elseif(data.response.ok===true){return20}else{return1}}
The one caveat with sampling and distributed tracing is that if you sample in both/all systems independently of one another, that itbecomes likely that you will get partial traces.
For Cloudflare Pages support, check out theHoneycomb Pages Plugin.
About
A library to easily send request events and traces tohttps://honeycomb.io
Topics
Resources
License
Code of conduct
Contributing
Security policy
Uh oh!
There was an error while loading.Please reload this page.
Stars
Watchers
Forks
Uh oh!
There was an error while loading.Please reload this page.
Contributors8
Uh oh!
There was an error while loading.Please reload this page.
