-
Notifications
You must be signed in to change notification settings - Fork 428
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Same stanza id for peers #3376
Same stanza id for peers #3376
Conversation
This comment has been minimized.
This comment has been minimized.
7dbe0fd
to
ccd2749
Compare
Codecov Report
@@ Coverage Diff @@
## master #3376 +/- ##
==========================================
+ Coverage 80.37% 80.71% +0.34%
==========================================
Files 397 397
Lines 32438 32445 +7
==========================================
+ Hits 26072 26188 +116
+ Misses 6366 6257 -109
Continue to review full report at Codecov.
|
This comment has been minimized.
This comment has been minimized.
Also test that pm keys are not accepted in muc config, but accepted correctly in the pm section.
ccd2749
to
b816c7d
Compare
small_tests_24 / small_tests / b816c7d internal_mnesia_24 / internal_mnesia / b816c7d small_tests_23 / small_tests / b816c7d dynamic_domains_pgsql_mnesia_24 / pgsql_mnesia / b816c7d dynamic_domains_mysql_redis_24 / mysql_redis / b816c7d mam_SUITE:rdbms_async_pool_prefs_cases:messages_filtered_when_prefs_default_policy_is_roster{error,{test_case_failed,"ASSERT EQUAL\n\tExpected []\n\tValue [ok]\n"}} ldap_mnesia_24 / ldap_mnesia / b816c7d dynamic_domains_mssql_mnesia_24 / odbc_mssql_mnesia / b816c7d dynamic_domains_pgsql_mnesia_23 / pgsql_mnesia / b816c7d ldap_mnesia_23 / ldap_mnesia / b816c7d elasticsearch_and_cassandra_24 / elasticsearch_and_cassandra_mnesia / b816c7d pgsql_mnesia_24 / pgsql_mnesia / b816c7d mssql_mnesia_24 / odbc_mssql_mnesia / b816c7d inbox_SUITE:one_to_one:user_has_two_conversations{error,
{timeout_when_waiting_for_stanza,
[{escalus_client,wait_for_stanza,
[{client,
<<"kate_user_has_two_conversations_3.258357@localhost/res1">>,
escalus_tcp,<0.17415.0>,
[{event_manager,<0.17319.0>},
{server,<<"localhost">>},
{username,<<"kate_user_has_two_conversations_3.258357">>},
{resource,<<"res1">>}],
[{event_client,
[{event_manager,<0.17319.0>},
{server,<<"localhost">>},
{username,
<<"kate_user_has_two_conversations_3.258357">>},
{resource,<<"res1">>}]},
{resource,<<"res1">>},
{username,<<"kate_user_has_two_conversations_3.258357">>},
{server,<<"localhost">>},
{host,<<"localhost">>},
{port,5222},
{auth,{escalus_auth,auth_plain}},
{wspath,undefined},
{username,<<"kate_user_has_two_conversations_3.258357">>},
{server,<<"localhost">>},
{password,<<"makrowe;p">>},
{stream_id,<<"0bbe548dff5400ee">>}]},
5000],
[{file,
"/home/circleci/app/big_tests/_build/default/lib/escalus/src/escalus_client.erl"},
{line,136}]},
{inbox_helper,get_inbox,3,
[{file,"/home/circleci/app/big_tests/tests/inbox_helper.erl"},
{line,221}]},
{inbox_helper,check_inbox,4,
[{file,"/home/circleci/app/big_tests/tests/inbox_helper.erl"},
{line,167}]},
{inbox_SUI... pgsql_mnesia_23 / pgsql_mnesia / b816c7d mysql_redis_24 / mysql_redis / b816c7d riak_mnesia_24 / riak_mnesia / b816c7d mssql_mnesia_24 / odbc_mssql_mnesia / b816c7d service_domain_db_SUITE:db:db_keeps_syncing_after_cluster_join{error,{test_case_failed,{[<<"example1.com">>],
[<<"example1.com">>,<<"example2.com">>]}}} service_domain_db_SUITE:db:rest_with_auth:rest_delete_domain_cleans_data_from_mam{error,
{timeout_when_waiting_for_stanza,
[{escalus_client,wait_for_stanza,
[{client,
<<"bob_rest_delete_domain_cleans_data_from_mam_47.51066@example.org/res1">>,
escalus_tcp,<0.13073.2>,
[{event_manager,<0.13067.2>},
{server,<<"example.org">>},
{username,
<<"bob_rest_delete_domain_cleans_data_from_mam_47.51066">>},
{resource,<<"res1">>}],
[{event_client,
[{event_manager,<0.13067.2>},
{server,<<"example.org">>},
{username,
<<"bob_rest_delete_domain_cleans_data_from_mam_47.51066">>},
{resource,<<"res1">>}]},
{resource,<<"res1">>},
{username,
<<"bob_rest_delete_domain_cleans_data_from_mam_47.51066">>},
{server,<<"example.org">>},
{host,<<"localhost">>},
{port,5232},
{auth,{escalus_auth,auth_plain}},
{wspath,undefined},
{username,
<<"bob_rest_delete_domain_cleans_data_from_mam_47.51066">>},
{server,<<"example.org">>},
{host,<<"localhost">>},
{password,<<"makota3">>},
{port,5232},
{stream_id,<<"36c3a41c304a6ef5">>}]},
5000],
[{file,
"/home/circleci/app/big_tests/_build/default/lib/escalus/src/escalus_client.erl"},
{line,136}]},
{service_domain_db_SUITE,
'-rest_delete_domain_cleans_data_from_mam/1-fun-0-',5,
... service_domain_db_SUITE:db:rest_without_auth:rest_delete_domain_cleans_data_from_mam{error,
{timeout_when_waiting_for_stanza,
[{escalus_client,wait_for_stanza,
[{client,
<<"bob_rest_delete_domain_cleans_data_from_mam_55.514756@example.org/res1">>,
escalus_tcp,<0.13712.2>,
[{event_manager,<0.13706.2>},
{server,<<"example.org">>},
{username,
<<"bob_rest_delete_domain_cleans_data_from_mam_55.514756">>},
{resource,<<"res1">>}],
[{event_client,
[{event_manager,<0.13706.2>},
{server,<<"example.org">>},
{username,
<<"bob_rest_delete_domain_cleans_data_from_mam_55.514756">>},
{resource,<<"res1">>}]},
{resource,<<"res1">>},
{username,
<<"bob_rest_delete_domain_cleans_data_from_mam_55.514756">>},
{server,<<"example.org">>},
{host,<<"localhost">>},
{port,5232},
{auth,{escalus_auth,auth_plain}},
{wspath,undefined},
{username,
<<"bob_rest_delete_domain_cleans_data_from_mam_55.514756">>},
{server,<<"example.org">>},
{host,<<"localhost">>},
{password,<<"makota3">>},
{port,5232},
{stream_id,<<"0e830232de39517d">>}]},
5000],
[{file,
"/home/circleci/app/big_tests/_build/default/lib/escalus/src/escalus_client.erl"},
{line,136}]},
{service_domain_db_SUITE,
'-rest_delete_domain_cleans_data_from_mam/1-fun-0-',5... mssql_mnesia_24 / odbc_mssql_mnesia / b816c7d |
Implement a configuration flag to allow for the mam_id to be shared between sender and receiver in 1:1 chats. This works simply by setting the mam_id as persistent in the mongoose_acc, so that the receiver, if on the same cluster, can reuse the same ID.