-
Notifications
You must be signed in to change notification settings - Fork 464
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Data Week: Internal Batch Scoring MBD API Endpoint (Part 2) #2680
Data Week: Internal Batch Scoring MBD API Endpoint (Part 2) #2680
Comments
The info for the task can be found in the new BatchModelScoringRequest object. You can query the db for objects withe a status PENDING. I didn't add an IN-PROGRESS status, but maybe we should. The record includes a filename field. This is the name of the address list file. It can be found in S3 at (the The record also includes a Once complete, the results should be saved to The bucket and folder names are saved as settings in the Django app. |
@erichfi @stefi-says @NadjibBenlaldj Here is the documentation on how to upload, process, and access bulk processing of addresses https://github.com/passportxyz/passport-scorer/blob/main/api/registry/bulk-mbd-analysis.md. Just lmk if you have any question |
User Story:
As a data engineer, I want to set up an internal batch scoring MBD API endpoint, so that I can process large datasets efficiently for the data team and provide results in a downloadable CSV file.
Acceptance Criteria:
GIVEN the internal API endpoint,
WHEN the data team submits a list of addresses with their API key,
THEN the API should provide an estimated processing time and a job ID, allow status checks via a separate endpoint, and return an S3 bucket link to download the CSV file with the results when the job is completed.
Tech Details:
The rest of the story is covered in part 2
Open Questions:
Notes/Assumptions:
Ensure the endpoint can handle large datasets efficiently.
Open Questions:
Notes/Assumptions:
The text was updated successfully, but these errors were encountered: