- Notifications
You must be signed in to change notification settings - Fork914
chore: route connection logs to new table#18340
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to ourterms of service andprivacy statement. We’ll occasionally send you account related emails.
Already on GitHub?Sign in to your account
base:main
Are you sure you want to change the base?
Conversation
This comment was marked as resolved.
This comment was marked as resolved.
Sorry, something went wrong.
Uh oh!
There was an error while loading.Please reload this page.
Requesting a draft review just to confirm we're happy with the DB schema, the RBAC setup, and the overall direction. |
Uh oh!
There was an error while loading.Please reload this page.
Uh oh!
There was an error while loading.Please reload this page.
Uh oh!
There was an error while loading.Please reload this page.
Uh oh!
There was an error while loading.Please reload this page.
Uh oh!
There was an error while loading.Please reload this page.
Uh oh!
There was an error while loading.Please reload this page.
Uh oh!
There was an error while loading.Please reload this page.
Uh oh!
There was an error while loading.Please reload this page.
Uh oh!
There was an error while loading.Please reload this page.
Uh oh!
There was an error while loading.Please reload this page.
Uh oh!
There was an error while loading.Please reload this page.
Uh oh!
There was an error while loading.Please reload this page.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others.Learn more.
Nice work! I think this looks great in general, but the upsert part of the logic here doesn't fully make sense to me (see related comment). Could you elaborate on the intent?
return nil | ||
} | ||
func NewMock() *MockConnectionLogger { |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others.Learn more.
Suggestion: This feels more like a "fake" than a "mock".
connection_logs | ||
LEFT JOIN users ON | ||
connection_logs.user_id = users.id | ||
LEFT JOIN users as workspace_owner ON |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others.Learn more.
LEFT JOIN usersas workspace_ownerON | |
LEFT JOIN usersAS workspace_ownerON |
-- limit of 100 to prevent accidental excessively large queries. | ||
COALESCE(NULLIF(@limit_opt :: int, 0), 100) | ||
OFFSET | ||
@offset_opt; |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others.Learn more.
Note: Using limit/offset will get increasingly more slow as the table grows and you paginate further. First page will usually remain fast though.
Fine for now, though. Just raising awareness.
close_reason | ||
) VALUES | ||
($1, $2, $3, $4, $5, $6, $7, $8, $9, $10, $11, $12, $13, $14, $15) | ||
ON CONFLICT (connection_id, workspace_id, agent_name) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others.Learn more.
Considering we have both connect and disconnect enums andtype
isnot part of this conflict handling, it suggests to me there is some gap in the logic. Considering connect/disconnect both share the same connection ID, we'll never write the disconnect but upsert the connect with close time and reason.
Either writing two entries, or doing a upsert, might make sense. But I think we have to ask what are we trying to achieve here and how are we looking to display the data?
My take as a user would be that I want to see two separate entries in the log, appearing in their chronological order. When viewing the connect, it would be an additional nice bonus to be able to see when the disconnect happened, and vice-versa that could simply be achieved by applying the connection ID as search filter.
EDIT: Correction, I was referring to "connection_action" (connect/disconnect), not "connection_type". But upon closer inspection that database type is not even used so it should probably be removed if not intended to be used.
That said, I want to clarify that I'm not concerned with whether or not we write two entries or upsert one, but in the latter case I do think we should make sure to "explode" the data when queried so that we can show the events in their chronological order.
ethanndicksonJun 16, 2025 • edited
Loading Uh oh!
There was an error while loading.Please reload this page.
edited
Uh oh!
There was an error while loading.Please reload this page.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others.Learn more.
EDIT: Correction, I was referring to "connection_action" (connect/disconnect), not "connection_type". But upon closer inspection that database type is not even used so it should probably be removed if not intended to be used.
Part of the reason forconnection_action
is to better differentiateconnection_logs
when they're output to thecoder server
logs (the slog backend for the connection logger). Without including it on theInsertParams
, for a connect + disconnect, you'd just see two logs, one withclose_time
set, one without. The other reason is to ensure we don't incorrectly mark a connection as closed if an agent erroneously sent multipleconnect
events with the same connection ID. I do agree to see it's weird to see a database type only used in a query and not in the database, but this seemed like the most appropriate way to handle both these cases.
I do think we should make sure to "explode" the data when queried so that we can show the events in their chronological order.
This is a good point. In my head I envisaged the final UI just including a duration on the table, the idea being that theconnection
events were more important to the end user. Based off what I've heard of how users request / use this existing connection log data, I didn't think we'd be breaking any existing usecases by not showing disconnections intertwined with connections. Whilst I think we can explode the data in the query (and show connections and disconnection events chronologically), I'm not sure it'll play nice with pagination; i.e. you ask for offset 25, but instead of starting at the 26th disconnection or connection, your page starts after the most recent 25 connections.
Uh oh!
There was an error while loading.Please reload this page.
This is the first PR of a few for moving connection events out of the audit log, and into a new database table and web UI page called the 'Connection Log'.
This PR:
ConnectionLogger
abstraction to replace theAuditor
abstraction for these logs. (No-op'd in AGPL, like theAuditor
)ConnectionLogger
ConnectionLogger
instead of theAuditor
.Future PRs:
dbpurge
.Note
The PRs in this stack obviously won't be (completely) atomic. Whilst they'll each pass CI, the stack is designed to be merged all at once. I'm splitting them up for the sake of those reviewing, and so changes can be reviewed as early as possible. Despite this, it's really hard to make this PR any smaller than it already is. I'll be keeping it in draft until it's actually ready to merge.