Try to develop a spark program that performs the following:
1. run a "Pi Estimation" program in PySpark [login to view URL] and provide a screenshot of the results with a detailed explanation for how did you run the program
2. Please implement a Map/Reduce program on PySpark on the [login to view URL] attached file.
For each word in the input text, please compute the occurrences of other words that appear in every line that the word appeared. For example, assume that we have the following line as an input text. You are going to compute the occurrences of every word for all lines, not just for a single line.
An example of the required output for the [login to view URL] is on [login to view URL] in the attachments.
You must follow these guidelines to obtain the correct result.
Convert every string into a lower-case.
Remove all punctuation, i.e., any character other than a to z and space. It will be helpful to use a regular expression to replace them with empty strings. Of course, this will result in some non-sense situations (im, you're, and etc.), but you do not have to worry about this.
Please write a readme file showing how to run your code. Please Zip all files (including the source code)
Hi,
I have more that 6+years of experience in Hadoop and data mining technologies like HDFS, MapReduce, python, pypark, Scala, Hive etc.
Please review my profile for skills.
contact me.
$200 USD em 1 dia
4,8 (15 avaliações)
4,6
4,6
3 freelancers estão ofertando em média $127 USD for esse trabalho
hi there. i have read your description about the spark program. i can make it for you which includes both pi estimation and map reduce. here is about myself i am a professional python developer having expertise in data science machine learning , web development etc. lets connect we will discuss more about project, price, time frame in the chat section.
thank you.