About 8,460,000 results
Open links in new tab
  1. Pyspark: explode json in column to multiple columns

    Jun 28, 2018 · Pyspark: explode json in column to multiple columns Asked 7 years ago Modified 4 months ago Viewed 87k times

  2. pyspark - Adding a dataframe to an existing delta table throws …

    Jun 9, 2024 · You'll need to complete a few actions and gain 15 reputation points before being able to upvote. Upvoting indicates when questions and answers are useful. What's reputation and how do I get it? Instead, you can save this post to reference later.

  3. pyspark - How to use AND or OR condition in when in Spark

    105 pyspark.sql.functions.when takes a Boolean Column as its condition. When using PySpark, it's often useful to think "Column Expression" when you read "Column". Logical operations on PySpark columns use the bitwise operators: & for and | for or ~ for not When combining these with comparison operators such as <, parenthesis are often needed.

  4. Comparison operator in PySpark (not equal/ !=) - Stack Overflow

    Aug 24, 2016 · Comparison operator in PySpark (not equal/ !=) Asked 8 years, 11 months ago Modified 1 year, 5 months ago Viewed 164k times

  5. Pyspark: Parse a column of json strings - Stack Overflow

    I have a pyspark dataframe consisting of one column, called json, where each row is a unicode string of json. I'd like to parse each row and return a new dataframe where each row is the parsed json...

  6. PySpark: How to fillna values in dataframe for specific columns?

    Jul 12, 2017 · PySpark: How to fillna values in dataframe for specific columns? Asked 8 years ago Modified 6 years, 3 months ago Viewed 201k times

  7. How to save data frame in ".txt" file using pyspark

    Mar 23, 2018 · How to save data frame in ".txt" file using pyspark Asked 7 years, 4 months ago Modified 1 year, 11 months ago Viewed 71k times

  8. Pyspark : Dynamically prepare pyspark-sql query using parameters

    Sep 25, 2019 · What are the different ways to dynamicaly bind parameters and prepare pyspark-sql statament. Example: Dynamic Query query = '''SELECT column1, column2 FROM ${db_name}.${table_name} ...

  9. Show distinct column values in pyspark dataframe - Stack Overflow

    With pyspark dataframe, how do you do the equivalent of Pandas df['col'].unique(). I want to list out all the unique values in a pyspark dataframe column. Not the SQL type way (registertemplate the...

  10. pyspark dataframe filter or include based on list

    Nov 4, 2016 · I am trying to filter a dataframe in pyspark using a list. I want to either filter based on the list or include only those records with a value in the list. My code below does not work: # define a