Pyspark- Take total sum of a column and use the value to divide another column












0















I have a dataframe df



>>> df = spark.createDataFrame([[1,0], [2,1], [3,1], [4,0], [5,1]], ['a', 'b'])
>>> df.show()
+---+---+
| a| b|
+---+---+
| 1| 0|
| 2| 1|
| 3| 1|
| 4| 0|
| 5| 1|
+---+---+


and



>>> nrows = df.count()


Using df, I created a new dataframe a that is an aggregate of df.



>>> a = df.groupby('b').count()
>>> a.show()
+---+-----+
| b|count|
+---+-----+
| 0| 2|
| 1| 3|
+---+-----+


I need to create a new column in a called ev. The value of ev on the ith row is given by



enter image description here



This is the output I'm expecting



+---+-----+------------------+
| b|count| ev_norm|
+---+-----+------------------+
| 0| 2| 1.25|
| 1| 3|0.8333333333333334|
+---+-----+------------------+


But I reached here by first, creating a new column num for the numerator (nrows-count) which yields 3,2 on the two rows. I then proceeded to calculate the denominator (denom=0.48) which is the same for all rows. Lastly, I created a new column using .withColumn that divides the num column with denom.



Is there a way I can perform the calculation of denom on the fly without having to precalculate it and perform the above operations in a single step?



I tried something like



a = a.withColumn('ev_norm', (nrows - F.col('count'))/F.sum(F.col('count')*(nrows - F.col('count'))))`


and



a = a.withColumn('ev_norm', (sum([F.col('count')*(nrows-F.col('count'))]))


but it either gives me an error or the wrong answer.










share|improve this question























  • You should notice that if you can do this computation on the fly, you will do it for each row and it has a high computational cost. Hence, it would be plausible to precompute this single value!

    – OmG
    Nov 27 '18 at 7:44











  • Don't post pictures of code/data/formulas.

    – pault
    Nov 27 '18 at 14:38
















0















I have a dataframe df



>>> df = spark.createDataFrame([[1,0], [2,1], [3,1], [4,0], [5,1]], ['a', 'b'])
>>> df.show()
+---+---+
| a| b|
+---+---+
| 1| 0|
| 2| 1|
| 3| 1|
| 4| 0|
| 5| 1|
+---+---+


and



>>> nrows = df.count()


Using df, I created a new dataframe a that is an aggregate of df.



>>> a = df.groupby('b').count()
>>> a.show()
+---+-----+
| b|count|
+---+-----+
| 0| 2|
| 1| 3|
+---+-----+


I need to create a new column in a called ev. The value of ev on the ith row is given by



enter image description here



This is the output I'm expecting



+---+-----+------------------+
| b|count| ev_norm|
+---+-----+------------------+
| 0| 2| 1.25|
| 1| 3|0.8333333333333334|
+---+-----+------------------+


But I reached here by first, creating a new column num for the numerator (nrows-count) which yields 3,2 on the two rows. I then proceeded to calculate the denominator (denom=0.48) which is the same for all rows. Lastly, I created a new column using .withColumn that divides the num column with denom.



Is there a way I can perform the calculation of denom on the fly without having to precalculate it and perform the above operations in a single step?



I tried something like



a = a.withColumn('ev_norm', (nrows - F.col('count'))/F.sum(F.col('count')*(nrows - F.col('count'))))`


and



a = a.withColumn('ev_norm', (sum([F.col('count')*(nrows-F.col('count'))]))


but it either gives me an error or the wrong answer.










share|improve this question























  • You should notice that if you can do this computation on the fly, you will do it for each row and it has a high computational cost. Hence, it would be plausible to precompute this single value!

    – OmG
    Nov 27 '18 at 7:44











  • Don't post pictures of code/data/formulas.

    – pault
    Nov 27 '18 at 14:38














0












0








0








I have a dataframe df



>>> df = spark.createDataFrame([[1,0], [2,1], [3,1], [4,0], [5,1]], ['a', 'b'])
>>> df.show()
+---+---+
| a| b|
+---+---+
| 1| 0|
| 2| 1|
| 3| 1|
| 4| 0|
| 5| 1|
+---+---+


and



>>> nrows = df.count()


Using df, I created a new dataframe a that is an aggregate of df.



>>> a = df.groupby('b').count()
>>> a.show()
+---+-----+
| b|count|
+---+-----+
| 0| 2|
| 1| 3|
+---+-----+


I need to create a new column in a called ev. The value of ev on the ith row is given by



enter image description here



This is the output I'm expecting



+---+-----+------------------+
| b|count| ev_norm|
+---+-----+------------------+
| 0| 2| 1.25|
| 1| 3|0.8333333333333334|
+---+-----+------------------+


But I reached here by first, creating a new column num for the numerator (nrows-count) which yields 3,2 on the two rows. I then proceeded to calculate the denominator (denom=0.48) which is the same for all rows. Lastly, I created a new column using .withColumn that divides the num column with denom.



Is there a way I can perform the calculation of denom on the fly without having to precalculate it and perform the above operations in a single step?



I tried something like



a = a.withColumn('ev_norm', (nrows - F.col('count'))/F.sum(F.col('count')*(nrows - F.col('count'))))`


and



a = a.withColumn('ev_norm', (sum([F.col('count')*(nrows-F.col('count'))]))


but it either gives me an error or the wrong answer.










share|improve this question














I have a dataframe df



>>> df = spark.createDataFrame([[1,0], [2,1], [3,1], [4,0], [5,1]], ['a', 'b'])
>>> df.show()
+---+---+
| a| b|
+---+---+
| 1| 0|
| 2| 1|
| 3| 1|
| 4| 0|
| 5| 1|
+---+---+


and



>>> nrows = df.count()


Using df, I created a new dataframe a that is an aggregate of df.



>>> a = df.groupby('b').count()
>>> a.show()
+---+-----+
| b|count|
+---+-----+
| 0| 2|
| 1| 3|
+---+-----+


I need to create a new column in a called ev. The value of ev on the ith row is given by



enter image description here



This is the output I'm expecting



+---+-----+------------------+
| b|count| ev_norm|
+---+-----+------------------+
| 0| 2| 1.25|
| 1| 3|0.8333333333333334|
+---+-----+------------------+


But I reached here by first, creating a new column num for the numerator (nrows-count) which yields 3,2 on the two rows. I then proceeded to calculate the denominator (denom=0.48) which is the same for all rows. Lastly, I created a new column using .withColumn that divides the num column with denom.



Is there a way I can perform the calculation of denom on the fly without having to precalculate it and perform the above operations in a single step?



I tried something like



a = a.withColumn('ev_norm', (nrows - F.col('count'))/F.sum(F.col('count')*(nrows - F.col('count'))))`


and



a = a.withColumn('ev_norm', (sum([F.col('count')*(nrows-F.col('count'))]))


but it either gives me an error or the wrong answer.







python pyspark apache-spark-sql






share|improve this question













share|improve this question











share|improve this question




share|improve this question










asked Nov 27 '18 at 6:40









Clock SlaveClock Slave

2,26852360




2,26852360













  • You should notice that if you can do this computation on the fly, you will do it for each row and it has a high computational cost. Hence, it would be plausible to precompute this single value!

    – OmG
    Nov 27 '18 at 7:44











  • Don't post pictures of code/data/formulas.

    – pault
    Nov 27 '18 at 14:38



















  • You should notice that if you can do this computation on the fly, you will do it for each row and it has a high computational cost. Hence, it would be plausible to precompute this single value!

    – OmG
    Nov 27 '18 at 7:44











  • Don't post pictures of code/data/formulas.

    – pault
    Nov 27 '18 at 14:38

















You should notice that if you can do this computation on the fly, you will do it for each row and it has a high computational cost. Hence, it would be plausible to precompute this single value!

– OmG
Nov 27 '18 at 7:44





You should notice that if you can do this computation on the fly, you will do it for each row and it has a high computational cost. Hence, it would be plausible to precompute this single value!

– OmG
Nov 27 '18 at 7:44













Don't post pictures of code/data/formulas.

– pault
Nov 27 '18 at 14:38





Don't post pictures of code/data/formulas.

– pault
Nov 27 '18 at 14:38












0






active

oldest

votes











Your Answer






StackExchange.ifUsing("editor", function () {
StackExchange.using("externalEditor", function () {
StackExchange.using("snippets", function () {
StackExchange.snippets.init();
});
});
}, "code-snippets");

StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "1"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);

StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});

function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});


}
});














draft saved

draft discarded


















StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53494078%2fpyspark-take-total-sum-of-a-column-and-use-the-value-to-divide-another-column%23new-answer', 'question_page');
}
);

Post as a guest















Required, but never shown

























0






active

oldest

votes








0






active

oldest

votes









active

oldest

votes






active

oldest

votes
















draft saved

draft discarded




















































Thanks for contributing an answer to Stack Overflow!


  • Please be sure to answer the question. Provide details and share your research!

But avoid



  • Asking for help, clarification, or responding to other answers.

  • Making statements based on opinion; back them up with references or personal experience.


To learn more, see our tips on writing great answers.




draft saved


draft discarded














StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fstackoverflow.com%2fquestions%2f53494078%2fpyspark-take-total-sum-of-a-column-and-use-the-value-to-divide-another-column%23new-answer', 'question_page');
}
);

Post as a guest















Required, but never shown





















































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown

































Required, but never shown














Required, but never shown












Required, but never shown







Required, but never shown







Popular posts from this blog

A CLEAN and SIMPLE way to add appendices to Table of Contents and bookmarks

Calculate evaluation metrics using cross_val_predict sklearn

Insert data from modal to MySQL (multiple modal on website)