Posted 5d ago

Associate- BIM (10507)

@ Axtria
Hyderabad, Telangana, India
OnsiteFull Time
Responsibilities:Write PySpark queries, Design ETL workflows, Write complex SQL queries
Requirements Summary:Be/B.Tech with MCA; 4+ years in Spark/PySpark data processing; strong PySpark, Python, SQL skills; ETL design; pharma/Life Sciences domain preferred.
Technical Tools Mentioned:PySpark, Python, Pandas, NumPy, SQL, Databricks, Snowflake
Save
Mark Applied
Hide Job
Report & Hide
Job Description

Career Opportunities: Associate- BIM (10507)

Requisition ID 10507 - Posted  - Hyderabad





































 


Position Summary

To be a driven business analyst who can work on complex Analytical problems and help the customer in better business decision making especially in the area of pharma/life sciences (domain).

Job Responsibilities

Write Pyspark queries for data transformation needs. 

Participates in ETL Design using any python framework of new or changing mappings and workflows with the team and prepares technical specifications  

Write complex SQL queries with performance tuning and optimization 

Should be able to handle task independently and lead the team if required.

Good communication Skills

Coordinate with cross-functional teams to ensure project objectives are met.

Collaborate with data architects and engineers to design and implement data models. 

Education

BE/B.Tech
Master of Computer Application

Work Experience

Advanced knowledge of PySpark ,python, pandas, numpy frameworks. 

Minimum 4 years of extensive experience in design, build and deployment of Spark/Pyspark - for data integration. 

Deep experience in developing data processing tasks using pySpark such as reading data from external sources, merge data, perform data enrichment and load in to target data destinations 

Create Spark jobs for data transformation and aggregation 

Spark query tuning and performance optimization - Good understanding of different file formats (ORC, Parquet, AVRO) to optimize queries/processing and compression techniques. 

Deep understanding of distributed systems (e.g. CAP theorem, partitioning, replication, consistency, and consensus) 

Experience in Modular Programming & Robust programming methodologies 

ETL knowledge and have done ETL development using any python framework

Worked with Databricks/Snowflake in the past Preferable. 

 

Behavioural Competencies

Ownership
Teamwork & Leadership
Cultural Fit
Motivation to Learn and Grow

Technical Competencies

Problem Solving
Lifescience Knowledge
Communication
Capability Building / Thought Leadership

Skills










 
































Email this job to a friend
 
 
 
The job has been sent to
 






Please provide the information below
Job title:
*Your friend’s email address:
Message:



*Confirm you are not a robot:






Requisition ID 10507 - Posted  - Hyderabad


Position Summary

To be a driven business analyst who can work on complex Analytical problems and help the customer in better business decision making especially in the area of pharma/life sciences (domain).

Job Responsibilities

Write Pyspark queries for data transformation needs. 

Participates in ETL Design using any python framework of new or changing mappings and workflows with the team and prepares technical specifications  

Write complex SQL queries with performance tuning and optimization 

Should be able to handle task independently and lead the team if required.

Good communication Skills

Coordinate with cross-functional teams to ensure project objectives are met.

Collaborate with data architects and engineers to design and implement data models. 

Education

BE/B.Tech
Master of Computer Application

Work Experience

Advanced knowledge of PySpark ,python, pandas, numpy frameworks. 

Minimum 4 years of extensive experience in design, build and deployment of Spark/Pyspark - for data integration. 

Deep experience in developing data processing tasks using pySpark such as reading data from external sources, merge data, perform data enrichment and load in to target data destinations 

Create Spark jobs for data transformation and aggregation 

Spark query tuning and performance optimization - Good understanding of different file formats (ORC, Parquet, AVRO) to optimize queries/processing and compression techniques. 

Deep understanding of distributed systems (e.g. CAP theorem, partitioning, replication, consistency, and consensus) 

Experience in Modular Programming & Robust programming methodologies 

ETL knowledge and have done ETL development using any python framework

Worked with Databricks/Snowflake in the past Preferable. 

 

Behavioural Competencies

Ownership
Teamwork & Leadership
Cultural Fit
Motivation to Learn and Grow

Technical Competencies

Problem Solving
Lifescience Knowledge
Communication
Capability Building / Thought Leadership

Skills



Email this job to a friend
 
The job has been sent to
 
The job has been sent to


Position Summary

To be a driven business analyst who can work on complex Analytical problems and help the customer in better business decision making especially in the area of pharma/life sciences (domain).

Job Responsibilities

Write Pyspark queries for data transformation needs. 

Participates in ETL Design using any python framework of new or changing mappings and workflows with the team and prepares technical specifications  

Write complex SQL queries with performance tuning and optimization 

Should be able to handle task independently and lead the team if required.

Good communication Skills

Coordinate with cross-functional teams to ensure project objectives are met.

Collaborate with data architects and engineers to design and implement data models. 

Education

BE/B.Tech
Master of Computer Application

Work Experience

Advanced knowledge of PySpark ,python, pandas, numpy frameworks. 

Minimum 4 years of extensive experience in design, build and deployment of Spark/Pyspark - for data integration. 

Deep experience in developing data processing tasks using pySpark such as reading data from external sources, merge data, perform data enrichment and load in to target data destinations 

Create Spark jobs for data transformation and aggregation 

Spark query tuning and performance optimization - Good understanding of different file formats (ORC, Parquet, AVRO) to optimize queries/processing and compression techniques. 

Deep understanding of distributed systems (e.g. CAP theorem, partitioning, replication, consistency, and consensus) 

Experience in Modular Programming & Robust programming methodologies 

ETL knowledge and have done ETL development using any python framework

Worked with Databricks/Snowflake in the past Preferable. 

 

Behavioural Competencies

Ownership
Teamwork & Leadership
Cultural Fit
Motivation to Learn and Grow

Technical Competencies

Problem Solving
Lifescience Knowledge
Communication
Capability Building / Thought Leadership

Skills