Discover Dollar is hiring an Analytics/AI Engineer Intern. This role focuses on designing and implementing robust, scalable data pipelines using PySpark, Python, Polars and Gen-AI; developing data models and transformation logic to support analytics and business intelligence; applying Object-Oriented Programming (OOPs) principles; and using Databricks and cloud platforms to process large datasets while ensuring high data quality.
Responsibilities and Technical Scope
- Design and implement robust, scalable data pipelines:
Build end-to-end pipelines using PySpark, Python, Polars and Gen-AI to reliably ingest, transform and deliver data for analytics and reporting. Emphasize scalability and robustness so pipelines handle growing volumes and evolving business needs.
- Develop data models and transformation logic:
Create and maintain data models and transformation logic that directly support analytics and business intelligence use cases. Ensure transformations produce clean, consistent outputs suitable for downstream analyses and dashboards.
- Leverage Python and OOPs principles:
Apply Python and Object-Oriented Programming (OOPs) concepts to design reusable, maintainable data tools and workflows. Structure code for readability, modularity and reuse across pipeline components.
- Utilize Databricks and cloud-based platforms:
Work with Databricks and cloud platforms to process and manage large datasets efficiently, integrating pipeline components and transformations into scalable environments.
- Collaborate across teams and ensure data quality:
Partner with data scientists, analysts and business stakeholders to gather requirements, deliver clean trusted data, and ensure high data quality and consistency across systems and reports.
Candidate Requirements and Work Culture Fit
- Strong Python programming skills:
Demonstrated command of Python programming to build and maintain data processing logic and automation.
- Solid Object-Oriented Programming (OOPs) understanding:
Good grasp of OOPs concepts to write structured, maintainable code for data tools and workflows.
- Advanced SQL expertise:
Proficiency in SQL for data manipulation, querying and validating transformed data supporting analytics needs.
- Experience with PySpark and Polars:
Hands-on experience using PySpark and Polars for distributed and in-memory data processing to handle various dataset sizes and performance requirements.
- Independent, fast-paced adaptability:
Ability to work independently in a fast-paced, agile environment while maintaining a problem-solving mindset and eagerness to learn and experiment.
- Communication and collaboration:
Strong communication and collaboration skills to align with stakeholders and ensure delivery of clean, trusted data products.
Conclusion
Discover Dollar’s Analytics/AI Engineer Intern role combines technical depth—PySpark, Python, Polars, Gen-AI, Databricks and advanced SQL—with collaborative data stewardship and OOPs-driven engineering. Candidates who bring strong Python and OOPs skills, experience in distributed and in-memory processing, a problem-solving mindset, and the ability to thrive in a fast-paced setting will fit this opportunity to build scalable, high-quality data solutions.