-
Notifications
You must be signed in to change notification settings - Fork 27
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Encoding issue when using spark with mesos #134
Comments
We need to ensure spark cluster encoding with UTF-8 Sorry this post is CN but we could see from its bottom comment on where we may need to modify. |
As for mesos, I put some ref I got from my phone for ensuring UTF-8 encoding(gpt-4) 😋
Remember, consistency across your entire deployment (drivers, executors, and the environment they run in) is key to ensuring that UTF-8 encoding is used throughout. |
Ok, it seems indeed that adding |
Great to know this is the line, we should add this in docs of spark connecter🫡 |
When using nebula-spark-connector data is not reaching the nebula storage with the right encoding.
The issue is only reproducing when spark is running in a mesos cluster.
By trying to do the following
The
value
arrives in nebula as??
More env data:
spark version: 3.0.2
java version: 1.8.0_392
mesos 1.11
I have tried multiple combinations of env:
(spark 3.2.1 or mesos 1.10) the results have not changed
Data should arrive in nebula as 测试 not as ??
(please note that running with spark local or even on kubernetes works fine.)
The text was updated successfully, but these errors were encountered: