- Overview
- Features and Functionality
- Subscription Breakdown
- Roles
- Architecture Diagram
- Getting Started
- API Documentation
- Examples
- Deployment
- Creating or debugging of SQL scripts with Flyway
- Monitoring and Logging
- Security & Quality Considerations
- Test Suite
- Contributing
- License
pip-subscription-management
is a microservice that deals with operations related to subscriptions, including all CRUD operations and the triggering of the fulfilment process. It does not do the fulfilment itself - this is the responsibility of the publication service.
It sits within the Court and Tribunal Hearings Service (CaTH hereafter), written with Spring Boot/Java.
For context, a subscription in CaTH is something a user can set up in order to be notified when a new publication is received. There are also third party subscriptions which are set up by administrators, which notify selected third party APIs.
In practice, the service is usually containerized within a hosted kubernetes environment within Azure.
All interactions with pip-subscription-management
are performed through the API (specified in API Documentation) either as a standalone service or via connections to other microservices.
- Creation and deletion (including bulk deletion) of subscriptions for a user or third party.
- Configuration of which list types a user would like to receive for a subscription.
- Ensures that subscriptions are only fulfilled if the user has permission to see the publication.
- Provides MI reporting endpoints, which are used to produce the MI report.
- Provides the ability to delete all subscriptions for a location. This is used when deleting a location from CaTH.
- The ability to retrieve all subscription channels, which is used in our Frontend when setting up third party subscriptions.
- Triggers a system admin and user email to be sent when subscriptions are deleted due to a location being deleted.
- Flyway for database modifications via SQL ingestion.
- Secure/Insecure Mode: Use of bearer tokens for authentication with the secure instance (if desired)
- OpenAPI Spec/Swagger-UI: Documents and allows users or developers to access API resources within the browser.
- Integration tests using TestContainers for dummy database operations.
Each subscription is associated with a user. The user can be a verified user (a person) or a third party.
A verified user can subscribe by case number, case URN or a location. It's worth noting that users can subscribe by Case Name in the frontend, however it is mapped to a number or URN in the database.
When a publication comes in via data management, it is checked to see if any subscriptions match any of the above subscription types.
If it does, and as long as the user has permission to see that publication, a notification is sent to publication services which will fulfil that subscription.
Third party subscriptions are set up by administrators. Third party's can only subscribe by List Type.
In this scenario, as long as the publication matches the list type and the third party has permission to see it, the notification will be sent to publication services to fulfil the subscription.
When an artefact is deleted, data management will inform subscription management.
If any third party subscriptions exist for the deleted artefact, subscription management will then trigger an empty artefact deletion request to publication services.
When a verified user subscribes by location, they also have the ability to configure which list types they would like to receive for that location.
If the user does not configure their list types, then by default all list types are taken into account (following the normal rules of permissions)
If the user does configure which list types they would like to receive, then this will be applied to all location subscriptions that the user has. There is no ability in the frontend for a user to configure list types per location, however if there was a need this service would handle it without any changes.
Each subscription has a channel. For verified users, this is always EMAIL.
For third parties, there can be a number of different API channels. This is selected by the administrator in the frontend when setting up a third party subscription.
Any endpoint that should require authentication, needs to be annotated either at controller or endpoint level with @IsAdmin.
The above diagram is somewhat simplified for readability (e.g. it does not include secure/insecure communications, but those are covered elsewhere).
- Java JDK 21 - this is used throughout all of our services.
- REST client of some description (e.g. Curl, Insomnia, Postman). Swagger-UI can also be used to send requests.
- Docker - used to run integration tests due to our use of TestContainers
- Azurite - Local Azure emulator used along with Azure Storage explorer for local storage.
- Azure Storage Explorer - Used for viewing and storing blobs within an Azurite instance locally.
- pip-dev-env - This repo provides a development environment wherein ensure all microservices, as well as external services (e.g. postgres & redis) are all running in tandem within the service. It eases the development process and is particularly helpful when working with cross-service communication, as it also reduces strain on local performance from having many separate IDE windows open.
- PostgreSQL - for local development, it will help to install Postgres. Ensure your postgres instance matches the relevant environment variables. Most devs on the project are just using this within a docker container.
- Some means of interfacing with the postgres database either locally or remotely. Good options include DataGrip, pgAdmin or psql. This will allow you to verify the impacts of your requests on the underlying database.
- Clone the repository
- Ensure all required environment variables have been set.
- Build using the command
./gradlew clean build
- Start the service using the command
./gradlew bootrun
in the newly created directory.
Environment variables are used by the service to control its behaviour in various ways.
These variables can be found within various separate CaTH Azure keyvaults. You may need to obtain access to this via a support ticket.
- Runtime secrets are stored in
pip-ss-{env}-kv
(where {env} is the environment where the given instance is running (e.g. production, staging, test, sandbox)). - Test secrets are stored in
pip-bootstrap-{env}-kv
with the same convention.
Python scripts to quickly grab all environment variables (subject to Azure permissions) are available for both runtime and test secrets.
Below is a table of currently used environment variables for starting the service, along with a descriptor of their purpose and whether they are optional or required.
Variable | Description | Required? |
---|---|---|
SPRING_PROFILES_ACTIVE | If set equal to dev , the application will run in insecure mode (i.e. no bearer token authentication required for incoming requests.) Note - if you wish to communicate with other services, you will need to set them all to run in insecure mode in the same way. |
No |
DB_HOST | Postgres Hostname | Yes |
DB_PORT | Postgres Port | Yes |
DB_NAME | Postgres Db name | Yes |
DB_USER | Postgres Username | Yes |
DB_PASS | Postgres Password | Yes |
APP_URI | Uniform Resource Identifier - the location where the application expects to receive bearer tokens after a successful authentication process. The application then validates received bearer tokens using the AUD parameter in the token | No |
CLIENT_ID | Unique ID for the application within Azure AD. Used to identify the application during authentication. | No |
TENANT_ID | Directory unique ID assigned to our Azure AD tenant. Represents the organisation that owns and manages the Azure AD instance. | No |
CLIENT_SECRET | Secret key for authentication requests to the service. | No |
ACCOUNT_MANAGEMENT_URL | URL used for connecting to the pip-account-management service. Defaults to staging if not provided. | No |
DATA_MANAGEMENT_URL | URL used for connecting to the pip-data-management service. Defaults to staging if not provided. | No |
PUBLICATION_SERVICES_URL | URL used for connecting to the pip-publication-services service. Defaults to staging if not provided. | No |
ACCOUNT_MANAGEMENT_AZ_API | Used as part of the scope parameter when requesting a token from Azure. Used for service-to-service communication with the pip-account management service |
No |
DATA_MANAGEMENT_AZ_API | Used as part of the scope parameter when requesting a token from Azure. Used for service-to-service communication with the pip-data-management service |
No |
PUBLICATION_SERVICES_AZ_API | Used as part of the scope parameter when requesting a token from Azure. Used for service-to-service communication with the pip-publication-services service |
No |
COURTEL_API | API value for third party | No |
ENABLE_TESTING_SUPPORT_API | Used to conditionally enable testing support API. Default to false for the production environment only. |
No |
Secrets required for getting tests to run correctly can be found in the below table:
Variable | Description |
---|---|
CLIENT_ID | Unique ID for the application within Azure AD. Used to identify the application during authentication. |
CLIENT_SECRET | Secret key for authentication requests to the service. |
DATA_MANAGEMENT_AZ_API | Used as part of the scope parameter when requesting a token from Azure. Used for service-to-service communication with the pip-data-management service. |
ACCOUNT_MANAGEMENT_AZ_API | Used as part of the scope parameter when requesting a token from Azure. Used for service-to-service communication with the pip-account-management service. |
PUBLICATION_SERVICES_AZ_API | Used as part of the scope parameter when requesting a token from Azure. Used for service-to-service communication with the pip-publication-services service. |
TENANT_ID | Directory unique ID assigned to our Azure AD tenant. Represents the organisation that owns and manages the Azure AD instance. |
SYSTEM_ADMIN_PROVENANCE_ID | Provenance ID for the test system admin user. |
CLIENT_ID_FT | Client ID of external service used for authentication with subscription-management application in the functional tests. |
CLIENT_SECRET_FT | Client secret of external service.used for authentication with subscription-management application in the functional tests. |
The service can also be adapted using the yaml files found in the following locations:
- src/main/resources/application.yaml for changes to the behaviour of the service itself.
- src/main/resources/application-dev.yaml for changes to the behaviour of the service when running locally.
- src/test/resources/application-test.yaml for changes to other test types (e.g. unit tests).
- src/integrationTest/resources/application-functional.yaml for changes to the application when its running functional tests.
We use Fortify to scan for security vulnerabilities. This is run as part of our nightly pipelines.
Our full API specification can be found within our Swagger-UI page. It can be accessed locally by starting the service and going to http://localhost:4550/swagger-ui/swagger-ui/index.html Alternatively, if you're on our VPN, you can access the swagger endpoint at our staging URL (ask a teammate to give you this).
As mentioned, the full api documentation can be found within swagger-ui, but some of the most common operations are highlighted below.
Most of the communication with this service benefits from using secure authentication. While possible to stand up locally in insecure mode, to simulate a production environment it is better to use secure mode. Before sending in any requests to the service, you'll need to obtain a bearer token using the following approach:
To request a bearer token, sending a post request following this template:
curl --request POST \
--url https://login.microsoftonline.com/{TENANT_ID}/oauth2/v2.0/token \
--header 'Content-Type: multipart/form-data' \
--form client_id={CLIENT_ID_FOR_ANOTHER_SERVICE} \
--form scope={APP_URI}/.default \
--form client_secret={CLIENT_SECRET_FOR_ANOTHER_SERVICE}\
--form grant_type=client_credentials
You can copy the above curl command into either Postman or Insomnia and they will automatically be converted to the relevant formats for those programs.
Note - the _FOR_ANOTHER_SERVICE
variables need to be extracted from another registered microservice within the broader CaTH umbrella (e.g. pip-data-management)
You can use the bearer token in the Authorization header when making requests. Here is an example using the create subscription endpoint.
curl --request POST \
--url http://localhost:4550/subscription \
--header 'Authorization: Bearer {BEARER_TOKEN_HERE}' \
--header 'Content-Type: application/json' \
--header 'x-user-id: <UserIDOfTheUser>' \
--data-raw '[
{
"userId": "<userId>",
"searchType": "CASE_ID",
"searchValue": "1234",
"channel": "EMAIL"
}
]'
We use Jenkins as our CI/CD system. The deployment of this can be controlled within our application logic using the various Jenkinsfile
-prepended files within the root directory of the repository.
Our builds run against our dev
environment during the Jenkins build process. As this is a microservice, the build process involves standing up the service in a docker container in a Kubernetes cluster with the current staging master copies of the other interconnected microservices.
If your debugging leads you to conclude that you need to implement a pipeline fix, this can be done in the CNP Jenkins repo
Flyway is used to apply incremental schema changes (migrations) to our database.
Any modifications to the database schema must be done through flyway. Changes to the models will no longer be automatically applied to the database.
This behaviour can be overridden using the DB_UPDATE environment variable. This is useful for local development, but should not be used in production.
Flyway is enabled on the pipeline. It is only run on the pipeline, and not on startup. On startup, the app will validate that the flyway scripts have been applied.
For local development, flyway is turned off by default. This is due to all tables existing within a single database locally. This can cause flyway to fail at startup due to mismatching scripts.
We utilise Azure Application Insights to store our logs. Ask a teammate for the specific resource in Azure to access these. Locally, we use Log4j.
In addition, this service is also monitored in production and staging environments by Dynatrace. The URL for viewing our specific Dynatrace instance can be had by asking a team member.
Application insights is configured via the lib/applicationinsights.json file. Alongside this, the Dockerfile is configured to copy in this file and also download the app insights client.
The client at runtime is attached as a javaagent, which allows it to send the logging to app insights.
To connect to app insights a connection string is used. This is configured to read from the KV Secret mounted inside the pod.
It is possible to connect to app insights locally, although somewhat tricky. The easiest way is to get the connection string from azure, set it as an environment variable (APPLICATIONINSIGHTS_CONNECTION_STRING), and add in the javaagent as VM argument. You will also need to remove / comment out the connection string line the config.
We use a few automated tools to ensure quality and security within the service. A few examples can be found below:
- SonarCloud - provides automated code analysis, finding vulnerabilities, bugs and code smells. Quality gates ensure that test coverage, code style and security are maintained where possible.
- DependencyCheckAggregate - Ensures that dependencies are kept up to date and that those with known security vulnerabilities (based on the National Vulnerability Database(NVD)) are flagged to developers for mitigation or suppression.
- JaCoCo Test Coverage - Produces code coverage metrics which allows developers to determine which lines of code are covered (or not) by unit testing. This also makes up one of SonarCloud's quality gates.
- PMD - Static code analysis tool providing code quality guidance and identifying potential issues relating to coding standards, performance or security.
- CheckStyle - Enforces coding standards and conventions such as formatting, naming conventions and structure.
This microservice is comprehensively tested using both unit and functional tests.
Unit tests can be run on demand using ./gradlew test
.
Functional tests can be run using ./gradlew functional
For our functional tests, we are using Square's MockWebServer library. This allows us to test the full HTTP stack for our service-to-service interactions.
The functional tests also call out to Data Management in staging to retrieve publications.
We are happy to accept third-party contributions. See .github/CONTRIBUTING.md for more details.
This project is licensed under the MIT License - see the LICENSE file for details