You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Hyukjin Kwon (JIRA)" <ji...@apache.org> on 2016/11/14 13:46:58 UTC

[jira] [Updated] (SPARK-18437) Inconsistent mark-down for `Note:` across Scala/Java/R/Python in API documentations

     [ https://issues.apache.org/jira/browse/SPARK-18437?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Hyukjin Kwon updated SPARK-18437:
---------------------------------
    Description: 
It seems in Scala/Java,

- {{Note:}}

- {{NOTE:}}

- {{Note that}}

- {{'''Note:'''}}

are used in a mixed way. The last one seems correct as it marks down correctly. This will, for example, mark down it pretty[1].

Also, it seems some {{'''Note:'''}}s are wrongly placed[2] which looks like the {{Note:}} for the last argument (I believe it meant to be for the API).

For Python, 

- {{Note:}}

- {{NOTE:}}

- {{Note that}}

- {{.. note:}}

In this case, I also believe the last one marks down pretty[3] rather than the others[4][5][6].

For R, it seems there are also,


- {{Note:}}

- {{NOTE:}}

- {{Note that}}

- {{@note}}

In case of R, it seems pretty consistent. {{@note}} only contains the information about when the function came out such as {{locate since 1.5.0}} without other information[7]. So, I am not too sure for this.

It would be nicer if they are consistent, at least for Scala/Python/Java.

[1]http://spark.apache.org/docs/latest/api/scala/index.html#org.apache.spark.SparkContext@hadoopFile[K,V,F<:org.apache.hadoop.mapred.InputFormat[K,V]](path:String)(implicitkm:scala.reflect.ClassTag[K],implicitvm:scala.reflect.ClassTag[V],implicitfm:scala.reflect.ClassTag[F]):org.apache.spark.rdd.RDD[(K,V)]
[2]http://spark.apache.org/docs/latest/api/scala/index.html#org.apache.spark.SparkContext@hadoopRDD[K,V](conf:org.apache.hadoop.mapred.JobConf,inputFormatClass:Class[_<:org.apache.hadoop.mapred.InputFormat[K,V]],keyClass:Class[K],valueClass:Class[V],minPartitions:Int):org.apache.spark.rdd.RDD[(K,V)]
[3]http://spark.apache.org/docs/latest/api/python/pyspark.sql.html#pyspark.sql.DataFrame.describe
[4]http://spark.apache.org/docs/latest/api/python/pyspark.sql.html#pyspark.sql.functions.date_format
[5]http://spark.apache.org/docs/latest/api/python/pyspark.sql.html#pyspark.sql.functions.grouping_id
[6]http://spark.apache.org/docs/latest/api/python/pyspark.sql.html#pyspark.sql.DataFrame.head
[7]http://spark.apache.org/docs/latest/api/R/index.html

  was:
It seems in Scala/Java,

- {{Note:}}

- {{NOTE:}}

- {{Note that}}

- {{'''Note:'''}}}

are used in a mixed way. The last one seems correct as it marks down correctly. This will, for example, mark down it pretty[1].

Also, it seems some {{'''Note:'''}}s are wrongly placed[2] which looks like the {{Note:}} for the last argument (I believe it meant to be for the API).

For Python, 

- {{Note:}}

- {{NOTE:}}

- {{Note that}}

- {{.. note:}}}

In this case, I also believe the last one marks down pretty[3] rather than the others[4][5][6].

For R, it seems there are also,


- {{Note:}}

- {{NOTE:}}

- {{Note that}}

- {{@note}}}

In case of R, it seems pretty consistent. {{@note}} only contains the information about when the function came out such as {{locate since 1.5.0}} without other information[7]. So, I am not too sure for this.

It would be nicer if they are consistent, at least for Scala/Python/Java.

[1]http://spark.apache.org/docs/latest/api/scala/index.html#org.apache.spark.SparkContext@hadoopFile[K,V,F<:org.apache.hadoop.mapred.InputFormat[K,V]](path:String)(implicitkm:scala.reflect.ClassTag[K],implicitvm:scala.reflect.ClassTag[V],implicitfm:scala.reflect.ClassTag[F]):org.apache.spark.rdd.RDD[(K,V)]
[2]http://spark.apache.org/docs/latest/api/scala/index.html#org.apache.spark.SparkContext@hadoopRDD[K,V](conf:org.apache.hadoop.mapred.JobConf,inputFormatClass:Class[_<:org.apache.hadoop.mapred.InputFormat[K,V]],keyClass:Class[K],valueClass:Class[V],minPartitions:Int):org.apache.spark.rdd.RDD[(K,V)]
[3]http://spark.apache.org/docs/latest/api/python/pyspark.sql.html#pyspark.sql.DataFrame.describe
[4]http://spark.apache.org/docs/latest/api/python/pyspark.sql.html#pyspark.sql.functions.date_format
[5]http://spark.apache.org/docs/latest/api/python/pyspark.sql.html#pyspark.sql.functions.grouping_id
[6]http://spark.apache.org/docs/latest/api/python/pyspark.sql.html#pyspark.sql.DataFrame.head
[7]http://spark.apache.org/docs/latest/api/R/index.html


> Inconsistent mark-down for `Note:` across Scala/Java/R/Python in API documentations
> -----------------------------------------------------------------------------------
>
>                 Key: SPARK-18437
>                 URL: https://issues.apache.org/jira/browse/SPARK-18437
>             Project: Spark
>          Issue Type: Improvement
>          Components: Documentation
>    Affects Versions: 2.0.1
>            Reporter: Hyukjin Kwon
>
> It seems in Scala/Java,
> - {{Note:}}
> - {{NOTE:}}
> - {{Note that}}
> - {{'''Note:'''}}
> are used in a mixed way. The last one seems correct as it marks down correctly. This will, for example, mark down it pretty[1].
> Also, it seems some {{'''Note:'''}}s are wrongly placed[2] which looks like the {{Note:}} for the last argument (I believe it meant to be for the API).
> For Python, 
> - {{Note:}}
> - {{NOTE:}}
> - {{Note that}}
> - {{.. note:}}
> In this case, I also believe the last one marks down pretty[3] rather than the others[4][5][6].
> For R, it seems there are also,
> - {{Note:}}
> - {{NOTE:}}
> - {{Note that}}
> - {{@note}}
> In case of R, it seems pretty consistent. {{@note}} only contains the information about when the function came out such as {{locate since 1.5.0}} without other information[7]. So, I am not too sure for this.
> It would be nicer if they are consistent, at least for Scala/Python/Java.
> [1]http://spark.apache.org/docs/latest/api/scala/index.html#org.apache.spark.SparkContext@hadoopFile[K,V,F<:org.apache.hadoop.mapred.InputFormat[K,V]](path:String)(implicitkm:scala.reflect.ClassTag[K],implicitvm:scala.reflect.ClassTag[V],implicitfm:scala.reflect.ClassTag[F]):org.apache.spark.rdd.RDD[(K,V)]
> [2]http://spark.apache.org/docs/latest/api/scala/index.html#org.apache.spark.SparkContext@hadoopRDD[K,V](conf:org.apache.hadoop.mapred.JobConf,inputFormatClass:Class[_<:org.apache.hadoop.mapred.InputFormat[K,V]],keyClass:Class[K],valueClass:Class[V],minPartitions:Int):org.apache.spark.rdd.RDD[(K,V)]
> [3]http://spark.apache.org/docs/latest/api/python/pyspark.sql.html#pyspark.sql.DataFrame.describe
> [4]http://spark.apache.org/docs/latest/api/python/pyspark.sql.html#pyspark.sql.functions.date_format
> [5]http://spark.apache.org/docs/latest/api/python/pyspark.sql.html#pyspark.sql.functions.grouping_id
> [6]http://spark.apache.org/docs/latest/api/python/pyspark.sql.html#pyspark.sql.DataFrame.head
> [7]http://spark.apache.org/docs/latest/api/R/index.html



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org