Backend/Data Engineer Intern
- $50k – $80k
- Remote •
- 3 years of exp
- Internship
Not Available
Remote only
About the job
About Us:
We are a fast-paced startup building the best AI Copilot for marketing teams. Our mission is to transform the marketing industry with AI-driven solutions that empower teams to work smarter, faster, and more effectively.
Team:
Aliveo AI is led by an expert team with deep experience in AI and machine learning. Joining us offers a rare opportunity to work alongside a top-tier team, contributing directly to cutting-edge AI products.
Job Description:
We are looking for a Backend/Data Engineer with strong expertise in Python, Flask, and data engineering to help us build and manage scalable data pipelines and systems that power our AI-driven marketing solutions. A critical aspect of the role will be working with data and integrating Large Language Models (LLMs), in addition to managing data from APIs such as Google Ads, Salesforce, HubSpot, Facebook, and others.
You will be responsible for designing and maintaining data pipelines, integrating new data sources, and helping deploy AI models, including LLMs, to optimize our product’s performance.
Key Responsibilities:
- Data Pipelines: Design, develop, and maintain robust data pipelines for collecting, transforming, and loading data from third-party APIs and internal sources.
- API Integration: Work with a range of APIs (e.g., Google Ads, Salesforce, HubSpot, Facebook) to pull in large volumes of data, ensuring seamless data flows into our systems.
- LLM Integration: Collaborate with team members to integrate and iterate on our Large Language Models (LLMs) data product.
- Backend Development: Build and maintain backend systems using Flask and PostgreSQL, ensuring scalability and performance for data-heavy applications.
- AWS Management: Implement and manage AWS services (EC2, S3, Lambda, RDS) for data processing and model deployment, optimizing cloud infrastructure for performance.
- Data Management: Organize and manage large datasets, optimizing them for real-time processing and analysis by data science and machine learning teams.
- Optimization: Troubleshoot, debug, and optimize backend systems, pipelines, and LLM integrations to ensure high performance and reliability.
- Collaboration: Work closely with cross-functional teams, including data scientists, machine learning engineers, and product teams to deliver data-driven features and improvements.
- Code Quality: Write clean, maintainable Python code, following best practices for both backend engineering and data pipeline development.
Internship Qualifications:
- Currently pursuing or recently completed a degree in Computer Science, Data Engineering, Software Engineering, or a related field.
- Self-motivated, proactive, and eager to learn and take on challenges.
- Demonstrated sense of urgency, ownership and eagerness to move fast and learn
- Prior experience working with Python ( Flask for backend development is a plus).
- Basic understanding of data pipelines and working with APIs (experience with platforms like Google Ads, Salesforce, HubSpot, or Facebook is a plus).
- Familiarity with SQL and databases (e.g., PostgreSQL).
- Exposure to or interest in Large Language Models (LLMs) and AI technologies.
- Willingness to learn AWS services (EC2, S3, Lambda, RDS) for deploying cloud-based systems.
- Strong problem-solving skills, communication skills and a passion for working with data.
- Good communication skills and the ability to collaborate effectively in a remote team environment.