15,909,503 members
Sign in
Sign in
Email
Password
Forgot your password?
Sign in with
home
articles
Browse Topics
>
Latest Articles
Top Articles
Posting/Update Guidelines
Article Help Forum
Submit an article or tip
Import GitHub Project
Import your Blog
quick answers
Q&A
Ask a Question
View Unanswered Questions
View All Questions
View C# questions
View C++ questions
View Javascript questions
View Visual Basic questions
View Python questions
discussions
forums
CodeProject.AI Server
All Message Boards...
Application Lifecycle
>
Running a Business
Sales / Marketing
Collaboration / Beta Testing
Work Issues
Design and Architecture
Artificial Intelligence
ASP.NET
JavaScript
Internet of Things
C / C++ / MFC
>
ATL / WTL / STL
Managed C++/CLI
C#
Free Tools
Objective-C and Swift
Database
Hardware & Devices
>
System Admin
Hosting and Servers
Java
Linux Programming
Python
.NET (Core and Framework)
Android
iOS
Mobile
WPF
Visual Basic
Web Development
Site Bugs / Suggestions
Spam and Abuse Watch
features
features
Competitions
News
The Insider Newsletter
The Daily Build Newsletter
Newsletter archive
Surveys
CodeProject Stuff
community
lounge
Who's Who
Most Valuable Professionals
The Lounge
The CodeProject Blog
Where I Am: Member Photos
The Insider News
The Weird & The Wonderful
help
?
What is 'CodeProject'?
General FAQ
Ask a Question
Bugs and Suggestions
Article Help Forum
About Us
Search within:
Articles
Quick Answers
Messages
Comments by Silpa Silpa (Top 2 by date)
Silpa Silpa
27-Aug-22 5:05am
View
Oh ok thank you for your confirmation.. i am not expecting some one to write full code for me.. i am using sql and i also know how to filter in sql but in this code they are asking about spark api i don't know how to filter using in spark api scala. That thing i don't know..so i posted full question for clarification only..i am just searching for example ..not for full code.. thankyou.. i think i reached wrong site...
Silpa Silpa
27-Aug-22 4:33am
View
: Spark Scala coding exercise based on below requirement.
Expectation is to implement all the requirements using latest SPARK API (not sql) and functionalities as per our cluster upgraded version (spark version - 2.4.0, scala -2.11.8).
Input:
1. Given below input Partner data in comma separated file. (file name - Partner.csv).
2. Given List of Invalid Party Ids.
Requirement:
1 Read the input Partner csv file and Invalid party ids list as Dataframes using spark APIs.
3. Retrieve the latest record per party id based on the Last updated Date from input Partner Data.
4. Retrieve the valid Party ids from the Latest record DataFrame based on lookup from Invalid Party DataFrame.