Eli Lilly and Company Logo

Eli Lilly and Company

Data Engineer (BU)

Reposted 4 Days Ago
Be an Early Applicant
In-Office
Hyderabad, Telangana
Mid level
In-Office
Hyderabad, Telangana
Mid level
The Data Engineer will design, develop, and maintain data solutions for analysis, including creating data pipelines, enhancing data quality, and collaborating with data scientists and stakeholders.
The summary above was generated by AI

At Lilly, we unite caring with discovery to make life better for people around the world. We are a global healthcare leader headquartered in Indianapolis, Indiana. Our employees around the world work to discover and bring life-changing medicines to those who need them, improve the understanding and management of disease, and give back to our communities through philanthropy and volunteerism. We give our best effort to our work, and we put people first. We’re looking for people who are determined to make life better for people around the world.

About the Tech@Lilly Organization:

Tech@Lilly builds and maintains capabilities using cutting edge technologies like most prominent tech companies. What differentiates Tech@Lilly is that we create new possibilities through tech to advance our purpose – creating medicines that make life better for people around the world, like data driven drug discovery and connected clinical trials. We hire the best technology professionals from a variety of backgrounds, so they can bring an assortment of knowledge, skills, and diverse thinking to deliver innovative solutions in every area of the enterprise.

About the Business Function:

Tech@Lilly Business Units is a global organization strategically positioned so that through information and technology leadership and solutions, we create meaningful connections and remarkable experiences, so people feel genuinely cared for. The Business Unit organization is accountable for designing, developing, and supporting commercial or customer engagement services and capabilities that span multiple Business Units (Bio-Medicines, Diabetes, Oncology, International), functions, geographies, and digital channels. The areas supported by Business Unit includes: Customer Operations, Marketing and Commercial Operations, Medical Affairs, Market Research, Pricing, Reimbursement and Access, Customer Support Programs, Digital Production and Distribution, Global Patient Outcomes, and Real-World Evidence.

Job Title: Data Engineer:

A Data Engineer is responsible for designing, developing, and maintaining the data solutions that ensure the availability and quality of data for analysis and/or business transactions. They design and implement efficient data storage, processing and retrieval solutions for datasets and build data pipelines, optimize database designs, and work closely with data scientists, architects, and analysts to ensure data quality and accessibility. Data engineers require strong skillsets in data integration, acquisition, cleansing, harmonization, and transforming data. They play a crucial role in transforming raw data into datasets designed for analysis which enable organizations to unlock valuable insights for decision making.

"This role is open across experience levels, and the final designation will be determined based on the interview and assessment outcomes."

What you’ll be doing:

  • Design, build, and maintain scalable and reliable data pipelines for batch and real-time processing.

  • Develop and optimize data models, ETL/ELT workflows, and data integration across multiple systems and platforms.

  • Collaborate with data scientists, analysts, and business stakeholders to understand data requirements and deliver solutions.

  • Implement data governance, security, and quality standards across data assets.

  • Lead end-to-end data engineering projects and contribute to architectural decisions.

  • Design and implement cloud-native solutions on AWS (preferred) using tools such as AWS Glue, EMR, and Databricks. Experience with Azure or GCP is a plus.

  • Promote best practices in coding, testing, and deployment.

  • Monitor, troubleshoot, and improve performance and reliability of data infrastructure.

  • Automate manual processes and identify opportunities to optimize data workflows and reduce costs.

How You Will Succeed: 

  • Deliver scalable solutions by designing robust data pipelines and architectures that meet performance and reliability standards.

  • Collaborate effectively with cross-functional teams to turn business needs into technical outcomes.

  • Lead with expertise, mentoring peers and driving adoption of best practices in data engineering and cloud technologies.

  • Continuously improve systems through automation, performance tuning, and proactive issue resolution.

  • Communicate with clarity to ensure alignment across technical and non-technical stakeholders.

What You should Bring:

  • Strong proficiency in SQL and Python.

  • Hands-on experience with cloud platforms (AWS, Azure, or GCP) and tools like Glue, EMR, Redshift, Lambda, or Databricks.

  • Deep understanding of ETL/ELT workflows, data modelling, and data warehousing concepts.

  • Familiarity with big data and streaming frameworks (e.g., Apache Spark, Kafka, Flink).

  • Knowledge of data governance, security, and quality practices.

  • Working knowledge of Databricks for building and optimizing scalable data pipelines and analytics workflows.

  • Experience with CI/CD, version control (Git), and infrastructure-as-code tools is a plus.

  • A problem-solving mindset, attention to detail, and a passion for clean, maintainable code.

  • Strong communication and collaboration skills to work with both technical and non-technical stakeholders.

Basic Qualifications and Experience Requirement:

  • Bachelor’s degree in Computer Science, Information Technology, Management Information Systems, or equivalent work experience.

  • Overall 5-12 years of experience in data engineering using core technologies such as SQL, Python, PySpark, and AWS services including Lambda, Glue, S3, Redshift, Athena, and IAM roles/policies.

  • 4+ years of experience working in Agile environments, with hands-on experience using GitHub and CI/CD pipelines for code deployment.

  • 4+ years of experience with orchestration tools like Airflow for workflow automation.

  • Proven experience in architecting and building high-performance, scalable data pipelines following Data Lakehouse, Data Warehouse, and Data Mart standards.

  • Strong expertise in data modelling, managing large datasets, and implementing secure, compliant data governance practices.

  • Hands-on experience with Databricks, including cluster management, workspace configuration, notebook development, and performance optimization.

  • Experience in leading a small team of data engineers and providing technical mentorship.

  • Ability to collaborate with business stakeholders to translate key business requirements into scalable technical solutions.

  • Familiarity with security models and developing solutions on large-scale, distributed data systems.

Additional Skills/Preferences:

  • Domain experience in healthcare, pharmaceutical ( Customer Master, Product Master, Alignment Master, Activity, Consent etc.  ), or regulated industries is a plus.

  • Partner with and influence vendor resources on solution development to ensure understanding of data and technical direction for solutions as well as delivery

  • AWS Certified Data Engineer - Associate

  • Databricks Certified Data Engineer (Associate or Professional)

  • Familiarity with AI/ML workflows and integrating machine learning models into data pipelines

Additional Information:

  • Total number of position open for this role : 20

Lilly is dedicated to helping individuals with disabilities to actively engage in the workforce, ensuring equal opportunities when vying for positions. If you require accommodation to submit a resume for a position at Lilly, please complete the accommodation request form (https://careers.lilly.com/us/en/workplace-accommodation) for further assistance. Please note this is for individuals to request an accommodation as part of the application process and any other correspondence will not receive a response.

Lilly does not discriminate on the basis of age, race, color, religion, gender, sexual orientation, gender identity, gender expression, national origin, protected veteran status, disability or any other legally protected status.

#WeAreLilly

Top Skills

Spark
AWS
Aws Glue
Azure
Ci/Cd
Databricks
Emr
Flink
GCP
Git
Kafka
Python
SQL

Similar Jobs

Yesterday
Remote or Hybrid
16 Locations
Senior level
Senior level
Cloud • Computer Vision • Information Technology • Sales • Security • Cybersecurity
The Sr. Software Engineer will create file format parsers, collaborate on machine learning features, and maintain software systems. Responsibilities include testing, optimization, and documentation.
Top Skills: AWSAzureBitbucketC++GCPGitJenkinsJIRAPythonRust
13 Days Ago
Remote or Hybrid
18 Locations
Senior level
Senior level
Cloud • Computer Vision • Information Technology • Sales • Security • Cybersecurity
The Engineering Manager will lead the Linux sensor development team, manage engineers, drive technical strategy, and ensure high code quality for cybersecurity features.
Top Skills: CC++EbpfKubernetesLinuxUnix
Yesterday
In-Office or Remote
6 Locations
Junior
Junior
Information Technology • Consulting
Manage the planning, development, and launch of websites on WordPress, ensuring client objectives are met. Collaborate with various teams for project success.
Top Skills: CSSHTMLJavaScriptPHPWordpress

What you need to know about the Hyderabad Tech Scene

Because of its proximity to leading research institutions and a government committed to the city's growth, Hyderabad's tech scene is booming. With plans to establish India's first "AI city," the city is on track to become one of the world's most anticipated tech hubs, with companies like TransUnion, Schrödinger and Freshworks, among others, already calling the city home.

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account