Job Description
Role :: Data Analyst/ Engineer (Python, Pyspark coding)
Location :: Pittsburgh, PA and Cleveland, OH
Key Responsibilities
- Analyze structured and unstructured data to identify trends, patterns, and business insights.
- Develop and maintain reports, dashboards, and KPIs for stakeholders.
- Write efficient Python scripts for data cleansing, transformation, and automation tasks.
- Use PySpark for processing large-scale datasets in distributed environments.
- Assist in ETL/ELT data pipeline development and monitoring.
- Collaborate with business users, data engineers, and analysts to gather requirements.
- Validate data quality and troubleshoot inconsistencies.
- Prepare documentation for data processes, workflows, and reports.
- Support ad-hoc data requests and analysis projects.
Required Skills
- Strong knowledge of
Python
(Pandas, NumPy, Matplotlib preferred).
- Hands-on experience with
PySpark
and Spark DataFrames.
- Good understanding of SQL queries, joins, views, and stored procedures.
- Familiarity with ETL processes and data warehousing concepts.
- Experience with Excel and reporting tools like Power BI / Tableau is a plus.
- Basic understanding of cloud platforms such as AWS / Azure is preferred.
- Strong analytical and problem-solving skills.
- Good communication and teamwork abilities.
Qualifications
- Bachelor’s degree in Computer Science, Information Systems, Statistics, or related field.
- 5-6 years of experience in Data Analysis / Data Engineering support roles.
- Banking or financial domain experience is a plus.
Looking for more opportunities?
Browse thousands of graduate jobs and entry-level positions.