Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Low-Code Migration QA] On-the-fly data comparison to QA connectors #22690

Closed
YowanR opened this issue Feb 10, 2023 · 5 comments
Closed

[Low-Code Migration QA] On-the-fly data comparison to QA connectors #22690

YowanR opened this issue Feb 10, 2023 · 5 comments

Comments

@YowanR
Copy link
Contributor

YowanR commented Feb 10, 2023

Based on the latest conversation around the QA topic, we would like to explore the idea of creating a tool that can help us compare the results between the low-code connector and the python connector without writing any data to storage. All the comparisons should happen in memory and the tool should create a report that highlights:

  1. The performance of each connector (Speed)
  2. If the records from the low-code connector matches those from the python connector (Accuracy).
@YowanR
Copy link
Contributor Author

YowanR commented Feb 10, 2023

@lazebnyi Can you take a look at this issue? The goal is to use this doc as the basis but we want to have a tool that ensures that no data is written to storage. Instead, all the data has to store in memory. Does this make sense?
If we can land this work, we should be able to unblock the QA process for our GA connectors.

@YowanR
Copy link
Contributor Author

YowanR commented Feb 10, 2023

cc'ing @evantahler as well for FYI (and also in case you have ideas to share on the topic)

@lazebnyi lazebnyi self-assigned this Feb 10, 2023
@evantahler
Copy link
Contributor

evantahler commented Feb 10, 2023

Well, if the goal is to use seeded sandbox data, and all streams are already tested via SAT/Connector Acceptance... isn't just running the acceptance tests enough? If expected_records match, then we are good?

If all the streams are not well seeded, perhaps the thing to do is to run the connector via docker and pipe the raw output to a file for both the old and new versions. Then, we can use a diff tool to see if there are any changes.

@YowanR
Copy link
Contributor Author

YowanR commented Feb 10, 2023

@evantahler This is a related but separate effort and I think both should happen in paralle. We're working on seeding the accounts but this effort allows us to use customer data for the QA effort without requiring an approval from the customers. I'm looping you in a doc and we can get back to this issue after we are aligned

@lazebnyi
Copy link
Collaborator

lazebnyi commented May 5, 2023

Done - #24421

@lazebnyi lazebnyi closed this as completed May 5, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

3 participants