External

Data Engineer

🏢 Aeris  •  📍 India

Sign up to view full application details.

Job Description

About Aeris Communications Inc. For more than three decades, Aeris has been a trusted cellular IoT leader enabling the biggest IoT programs and opportunities across Automotive, Utilities and Energy, Fleet Management and Logistics, Medical Devices, and Manufacturing. Our IoT technology expertise serves a global ecosystem of 7,000 enterprise customers and 30 mobile network operator partners, and 90 million IoT devices across the world. Aeris powers today’s connected smart world with innovative technologies and borderless connectivity that simplify management, enhance security, optimize performance, and drive growth. Job Summary As a Data Engineer, you will be responsible for designing, building, and maintaining scalable data pipelines and architectures. You will work closely with data scientists, analysts, and software engineers to ensure the availability, integrity, and quality of data for analytical and operational needs. Location - Noida, 5 days working on-site Key Responsibilities Design, develop, and maintain robust ETL pipelines to collect, process, and store data from multiple sources Build and optimize data models for analytics, reporting, and operational needs Implement best practices for data governance, data quality, and data security Collaborate with stakeholders to understand data requirements and translate them into technical solutions Integrate new data sources and expand the data platform capabilities Monitor and troubleshoot data workflows and infrastructure issues Maintain documentation related to data architecture, data sources, and data flows Minimum Requirements Bachelor’s or Master’s degree in Computer Science, Engineering, Information Technology, or related field 3–6 years of proven experience working as a Data Engineer. Strong proficiency in SQL,Python, or Scala Experience with big data technologies, such as Hadoop, Spark, Kafka, etc. Familiarity with cloud platforms (AWS, GCP, Azure) and cloud-native data solutions Experience with data warehousing (e.g., Redshift, Snowflake, BigQuery) Good understanding of ETL concepts, data modeling, and data architecture Experience with workflow orchestration tools (e.g., Airflow, Luigi) Excellent problem-solving skills and attention to detail Strong communication and collaboration skills Preferred Requirements Experience in real-time data processing and streaming Knowledge of machine learning workflows and MLOps Exposure to DevOps practices and CI/CD pipelines for data projects
View Full Description & Requirements →