diff --git a/docs/sparkr.md b/docs/sparkr.md
index a5235b2bf66a..dfa5278ef849 100644
--- a/docs/sparkr.md
+++ b/docs/sparkr.md
@@ -54,7 +54,7 @@ if (nchar(Sys.getenv("SPARK_HOME")) < 1) {
Sys.setenv(SPARK_HOME = "/home/spark")
}
library(SparkR, lib.loc = c(file.path(Sys.getenv("SPARK_HOME"), "R", "lib")))
-sc <- sparkR.session(master = "local[*]", sparkConfig = list(spark.driver.memory="2g"))
+sparkR.session(master = "local[*]", sparkConfig = list(spark.driver.memory = "2g"))
{% endhighlight %}
@@ -115,7 +115,7 @@ specifying `--packages` with `spark-submit` or `sparkR` commands, or if initiali
{% highlight r %}
-sc <- sparkR.session(sparkPackages = "com.databricks:spark-avro_2.11:3.0.0")
+sparkR.session(sparkPackages = "com.databricks:spark-avro_2.11:3.0.0")
{% endhighlight %}
diff --git a/examples/src/main/r/RSparkSQLExample.R b/examples/src/main/r/RSparkSQLExample.R
index f20875ca9efb..33e88e15fd47 100644
--- a/examples/src/main/r/RSparkSQLExample.R
+++ b/examples/src/main/r/RSparkSQLExample.R
@@ -144,7 +144,7 @@ write.df(df1, "data/test_table/key=1", "parquet", "overwrite")
write.df(df2, "data/test_table/key=2", "parquet", "overwrite")
# Read the partitioned table
-df3 <- read.df("data/test_table", "parquet", mergeSchema="true")
+df3 <- read.df("data/test_table", "parquet", mergeSchema = "true")
printSchema(df3)
# The final schema consists of all 3 columns in the Parquet files together