API and storage for logging event and audit messages, persisted in JSON within a PostgreSQL db, with a thin frontend provided by PostgREST
The logserver serves as the Audit Record Repository, as detailed in the Basic Audit Log Patterns (BALP) implementation guide.
- Event Schema
- Config
- Roles
- Access via JWT
- API use
- Advanced Query Examples
- Direct DB access
- Reporting tools
- Setup Development Environment
logserver is agnostic to the format, provided it is valid JSON. Any number
of database tables can be used, but only the single events
table is built
in, containing a PostgreSQL JSONB column, event
.
It is desirable to generate log events compliant with the FHIR audit event resource. To generate an AuditEvent
resource nested within each event
is cumbersome, however the following field
parity is recommended:
NB: all fields are considered optional unless marked as required
severity
: required element. Use built in log level rather than adding an additionalseverity
field:critical
: Critical condition with application. Includeemergency
andalert
levels in this category.error
: Error condition with the application.warning
: Warning needing attention before escalation to error.info
: Normal operational messages not requiring action. Includenotice
level in this category (normal but significant).debug
: Debug level messages, useful to application developers.
version
: required logserver schema version.action
: required element to describe the type of operation performed.create
: creating a new resource, such as adding a patient.read
: read/view/search - data retrieved or viewed w/o modification.update
: indicates existing data was modified.delete
: indicates data was removed or deleted.execute
: system or application function such as, program execution or perform a query/search.login
: specific category for the log-in actionlogout
: specific category for the log-out action
occurred
: Date-Time of the event, including timestamp information. This may duplicate the logging system timestamp (such asasctime
) but will always capture the time the event took place, not when it hit the logging server.subject
: the subject of the activity, i.e.Patient/ab-123-ef
agent
: actor involved in the event, generally the logged-in user:ip_address
: end user or requesting system's IP Addresstype
: i.e.system
oruser
who
: i.e.Practitioner/123-abc
source
: event reporter or system generating the audit event.observer
: base URL of the system generating the audit messagetype
: system type such asdhair2
or other predefined project category.version
: version of the observer (in contrast to top levelversion
)
entity
: data or objects useddetail
: list of tagged value pairs for conveying additional information. example pair might includeurl
:<full_url>
in contrast tosource.observer
query
: query parameters for query-type entities
outcome
: details in event of a failure or warning. useseverity
to capture level.
Copy default.env
to .env
and edit. Don't quote strings!
Role | Access |
---|---|
web_anon | read only access to the API |
event_logger | read / write access to the API |
For write access, generate a JWT using the generate_JWT.py
script in root
of project:
pip install pyjwt
JWT=`python3 generate_JWT.py`
or use any other approach signing with the configured value of PGRST_JWT_SECRET
(hint: see .env
) and the following payload:
{
"role": "event_logger"
}
- Save the
Encoded
JWT for use, passing as a bearer token:
export TOKEN=<JWT>
curl https://FQDN/events -X POST -H "Authorization: Bearer $TOKEN" \
-H "Content-Type: application/json" \
-d '{"event": {
"asctime": "2020-04-26 00:33:26,731",
"name": "sof_wrapper",
"levelname": "INFO",
"tags": [
"auth",
"login"
],
"subject": "Patient/41703",
"user": "Practitioner/SMART-1234"
}
}'
All events posted are available at:
/events
Sample query to fetch all events including login
in the list of tags
:
/events?event.tags=cs.{login}
Sample query to fetch all events where the patient
field equals "marcus aurelius"
:
/events?select=event&event-%3E%3Epatient=eq.marcus%20aurelius
Sample query to fetch events for given patient AND tag:
/events?select=event&event.tags=cs.{mme-calc}&event-%3E%3Epatient=eq.marcus%20aurelius
Fetch events based on datetime:
/events?select=event->>asctime,event->>message&event->>asctime=gte.2023-04-01&limit=25
Fetch events based on datetime, and include a number of isacc-specific fields:
/events?select=event-%3E%3Easctime,event-%3E%3Ename,event-%3E%3Emessage,event-%3E%3Eversion,event-%3E%3Elevelname,event-%3E%3ETwilio%20messages,event-%3E%3ECommunicationRequest&event-%3E%3Easctime=gte.2023-04-18&limit=25
See PostgREST API documentation for additional options
Example to find letstalktech
events filtered by source. Using
direct database access, look up available source
from the development logserver logs.inform.dev.cirg.uw.edu
:
select distinct(event->'source'->>'type') from events;
?column?
-----------------
dhair2/inform
shl-ltt-server
shl-ltt
external-client
Request all events of source->type
= shl-ltt-server
, from the dev letstalktech
logserver:
https://logs.inform.dev.cirg.uw.edu/events?select=event&event->source->>type=eq.shl-ltt-server
or the first 10 with source_type
= shl-ltt-server
since Jan 30, 2025:
or only the event->occurred
from the first 10 with source_type
= shl-ltt
since Jan 30, 2025:
To access the backing postgres database, invoke docker compose
from the
deployed directory as follows:
docker compose exec postgres psql postgres://app_user:secret@postgres:5432/app_db -c '\dt api.*'
To simplify queries, set api
as the search path, to make it the default schema:
SET search_path TO api;
These instructions detail how to set up a minimal logserver service, including steps to import a database from another logserver instance.
git clone git@github.com:uwcirg/logserver.git
# clone this repositorycp default.env .env
# follow instructions withindocker compose pull
# pulls the dependent imagesdocker compose up -d
# starts the containers
Test the install is functional. Obtain a JWT and use it to POST a simple example document. Correct the URL to fit.
pip install pyjwt
export TOKEN=`python3 generate_JWT.py`
export URL="http://unbiunium.cirg.washington.edu:3000/events"
curl $URL -X POST -H "Authorization: Bearer $TOKEN" -H "Content-Type: application/json" -d '{"event": {"name": "cmd line test"}}'
Browsing to the same URL should render the single message.
Obtain a logserver database export using the following command, from the deployed directory on the source system:
docker compose exec postgres bash -c '\
pg_dump \
--dbname app_db \
--no-owner \
--encoding utf8 \
--username=app_user ' > /tmp/logs.sql
Copy that export back to the target dev system and return to root directory of checkout from step #1
Halt the service, as dropping the database won't function while postgrest
maintains open connections. Then bring up only the database container
docker compose down
docker compose up -d postgres
Drop the existing database, necessary to prevent conflicts on import. Recreate an empty shell
docker compose exec postgres dropdb --username app_user app_db
docker compose exec postgres createdb --username app_user app_db
Import the pg_dump file from the source system
docker compose run -v /tmp/:/tmp/ postgres psql postgres://app_user:secret@postgres:5432/app_db --file /tmp/logs.sql
Finally, bring back up the full service, and confirm the URL above works
docker compose up -d