About 12,200,000 results
Open links in new tab
  1. pyspark - How to use AND or OR condition in when in Spark

    105 pyspark.sql.functions.when takes a Boolean Column as its condition. When using PySpark, it's often useful to think "Column Expression" when you read "Column". Logical operations on …

  2. Pyspark: explode json in column to multiple columns

    Jun 28, 2018 · Pyspark: explode json in column to multiple columns Asked 7 years ago Modified 3 months ago Viewed 86k times

  3. pyspark dataframe filter or include based on list

    Nov 4, 2016 · I am trying to filter a dataframe in pyspark using a list. I want to either filter based on the list or include only those records with a value in the list. My code below does not work: # …

  4. Pyspark: Parse a column of json strings - Stack Overflow

    I have a pyspark dataframe consisting of one column, called json, where each row is a unicode string of json. I'd like to parse each row and return a new dataframe where each row is the …

  5. PySpark: multiple conditions in when clause - Stack Overflow

    Jun 8, 2016 · when in pyspark multiple conditions can be built using & (for and) and | (for or). Note:In pyspark t is important to enclose every expressions within parenthesis () that combine …

  6. pyspark : NameError: name 'spark' is not defined

    Alternatively, you can use the pyspark shell where spark (the Spark session) as well as sc (the Spark context) are predefined (see also NameError: name 'spark' is not defined, how to solve?).

  7. spark dataframe drop duplicates and keep first - Stack Overflow

    Aug 1, 2016 · 2 I just did something perhaps similar to what you guys need, using drop_duplicates pyspark. Situation is this. I have 2 dataframes (coming from 2 files) which are exactly same …

  8. Manually create a pyspark dataframe - Stack Overflow

    Sep 16, 2019 · Manually create a pyspark dataframe Asked 5 years, 10 months ago Modified 1 year ago Viewed 208k times

  9. python - None/== vs Null/isNull in Pyspark? - Stack Overflow

    Jul 19, 2020 · Refer here : Filter Pyspark dataframe column with None value Equality based comparisons with NULL won't work because in SQL NULL is undefined so any attempt to …

  10. pyspark - Apache spark dealing with case statements - Stack …

    Oct 11, 2016 · I am dealing with transforming SQL code to PySpark code and came across some SQL statements. I don't know how to approach case statments in pyspark? I am planning on …