SHOULD YOU ACCEPT THIS CHALLENGE...
We are seeking a skilled and motivatedData Engineer and Backend Developerwith expertise in Python and proficiency in at least one other programming language. The ideal candidate should have hands-on experience in building ETL pipelines, manipulating structured and unstructured data, and optimizing data workflows. The role requires a strong understanding of databases, SQL, CSV, and JSON, as well as backend development for data processing and API integration.Experience in supply chain data or related domains will be a plus.
WHAT YOU'LL NEED TO BRING TO THIS ROLE...
Design, develop, and maintain scalable ETL pipelines to process and transform large data sets from multiple sources (e.g., CSV, JSON, databases).
Collaborate with data analysts and data scientists to ensure data is clean, accessible, and optimized for various data applications.
Develop backend systems and APIs to manage data ingestion, transformation, and storage efficiently.
Work with relational databases (SQL) to design, query, and optimize data storage solutions.
Write efficient, clean, and well-documented code in Python and other programming languages to support data processing and backend development tasks.
Ensure data security, integrity, and availability by implementing best practices in database management and ETL processes.
Collaborate with cross-functional teams to understand requirements and translate them into technical specifications and robust solutions.
Troubleshoot and debug data pipelines, systems, and applications as needed.
Stay updated on emerging trends in data engineering, backend development, and best practices in data management.
We are primarily an in-office environment and therefore, you will be expected to work from the Bangalore, India office in compliance with Pure's policies, unless you are on PTO, or work travel, or other approved leave.
Qualifications:
Bachelor's degree in Computer Science, Data Engineering, or a related field (or equivalent work experience).
Proficiency in Pythonandone or moreprogramming languages such as Java, C#, Ruby, Go, etc.
Strong understanding and hands-on experience withETL (Extract, Transform, Load)processes.
Solid experience withSQLdatabases and writing complex queries for data transformation and extraction.
Familiarity withfile formatssuch asCSV,JSON, and experience handling structured and unstructured data.
Experience withRESTful APIsand backend services for data integration.
Supply chain dataexperience or familiarity with supply chain management systems is a plus.
Familiarity with cloud platforms (e.g., AWS, Azure, Google Cloud) and distributed systems is a plus.
Experience with version control systems like Git and familiarity with CI/CD pipelines.
Strong analytical and problem-solving skills, with attention to detail.
Excellent communication skills and the ability to work both independently and collaboratively within a team.
Preferred Skills:
Experience with cloud-based data storage solutions such as AWS S3, Google Cloud Storage, or Azure Blob Storage.
Experience with NoSQL databases (e.g., MongoDB, Cassandra) is a plus.
Familiarity with big data tools and frameworks like Hadoop, Spark, or Kafka.
Experience in building data infrastructure for machine learning pipelines is an added advantage.
This role may be eligible for incentive pay and/or equity.
And because we understand the value of bringing your full and best self to work, we offer a variety of perks to manage a healthy balance, including flexible time off, wellness resources, and company-sponsored team events - check out purebenefits.com for more information.