-
Notifications
You must be signed in to change notification settings - Fork 55
/
test_automatic_fields.py
75 lines (58 loc) · 3.5 KB
/
test_automatic_fields.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
"""
Test that with no fields selected for a stream automatic fields are still replicated
"""
from tap_tester import runner, menagerie
from base import BaseTapTest
class MinimumSelectionTest(BaseTapTest):
"""Test that with no fields selected for a stream automatic fields are still replicated"""
@staticmethod
def name():
return "tt_jira_no_fields_test"
def test_run(self):
"""
- Verify that for each stream you can get multiple pages of data
- when no fields are selected and only the automatic fields are replicated.
- Verify that all replicated records have unique primary key values
PREREQUISITE
For EACH stream add enough data that you surpass the limit of a single
fetch of data. For instance if you have a limit of 250 records ensure
that 251 (or more) records have been posted for that stream.
"""
conn_id = self.create_connection_with_initial_discovery()
self.create_test_data()
# Select all streams and no fields within streams
# IF THERE ARE NO AUTOMATIC FIELDS FOR A STREAM
# WE WILL NEED TO UPDATE THE BELOW TO SELECT ONE
found_catalogs = menagerie.get_catalogs(conn_id)
expected_streams = self.expected_streams()
our_catalogs = [catalog for catalog in found_catalogs if
catalog.get('tap_stream_id') in expected_streams]
self.select_all_streams_and_fields(conn_id, our_catalogs, select_all_fields=False)
# Run a sync job using orchestrator
record_count_by_stream = self.run_sync(conn_id)
actual_fields_by_stream = runner.examine_target_output_for_fields()
synced_records = runner.get_records_from_target_output()
for stream in expected_streams:
with self.subTest(stream=stream):
# gather expectations
expected_primary_keys = self.expected_primary_keys().get(stream, set())
expected_automatic_fields = (expected_primary_keys |
self.top_level_replication_key_fields().get(stream, set()) |
self.expected_foreign_keys().get(stream, set()))
api_limit = self.expected_metadata().get(stream, {}).get(self.API_LIMIT)
# collect results
messages = synced_records.get(stream)
record_count = record_count_by_stream.get(stream, -1)
fields_replicated = actual_fields_by_stream.get(stream, set())
records_pks_list = [tuple([message.get('data').get(primary_key) for primary_key in expected_primary_keys])
for message in messages.get('messages')]
# verify that you get more than a page of data
self.assertGreater(record_count, api_limit,
logging="verify multiple pages are replicated")
# verify that only the automatic fields are sent to the target
self.assertEqual(fields_replicated, expected_automatic_fields,
logging="verify only automatic fields are replicated")
# Verify that all replicated records have unique primary key values
self.assertCountEqual(set(records_pks_list), records_pks_list,
msg="We have duplicate records for {}".format(stream),
logging="verify all records have unique primary key values")