diff --git a/docs/sql-programming-guide.md b/docs/sql-programming-guide.md
index c7ad06c63971..e59c327915c0 100644
--- a/docs/sql-programming-guide.md
+++ b/docs/sql-programming-guide.md
@@ -1851,7 +1851,8 @@ You can access them by doing
The value type in Scala of the data type of this field
(For example, Int for a StructField with the data type IntegerType) |
- StructField(name, dataType, nullable)
+ StructField(name, dataType, [nullable])
+ Note: The default value of nullable is true.
|
@@ -2139,7 +2140,8 @@ from pyspark.sql.types import *
The value type in Python of the data type of this field
(For example, Int for a StructField with the data type IntegerType) |
- StructField(name, dataType, nullable)
+ StructField(name, dataType, [nullable])
+ Note: The default value of nullable is True.
|
@@ -2260,7 +2262,7 @@ from pyspark.sql.types import *
vector or list |
list(type="array", elementType=elementType, containsNull=[containsNull])
- Note: The default value of containsNull is True.
+ Note: The default value of containsNull is TRUE.
|
@@ -2268,7 +2270,7 @@ from pyspark.sql.types import *
| environment |
list(type="map", keyType=keyType, valueType=valueType, valueContainsNull=[valueContainsNull])
- Note: The default value of valueContainsNull is True.
+ Note: The default value of valueContainsNull is TRUE.
|
@@ -2285,7 +2287,8 @@ from pyspark.sql.types import *
| The value type in R of the data type of this field
(For example, integer for a StructField with the data type IntegerType) |
- list(name=name, type=dataType, nullable=nullable)
+ list(name=name, type=dataType, nullable=[nullable])
+ Note: The default value of nullable is TRUE.
|