Rapsodo is a Sports Technology company with offices in the USA, Singapore, Turkey & Japan. We develop sports analytics products that are data-driven, portable and easy-to-use to empower athletes at all skill levels to analyse and improve their performance. From Major League Baseball star pitchers to Golf tour players, athletes use Rapsodo technology to up their game across the world. Trusted by coaches and players from youths to professionals, Rapsodo provides real-time insights for all-time performance.
We are innovative, focused, and rapidly growing. We are continuously looking for team players who will stop at nothing to deliver state-of-the-art solutions as part of Team Rapsodo.
About the role:
As a Data Engineering Intern at Rapsodo, you’ll play an essential role in building and optimizing the data infrastructure that powers decision-making and innovation throughout our organization. This role is ideal for students passionate about data engineering, with a foundational knowledge of data analytics to support meaningful insights. You’ll gain hands-on experience with cutting-edge technologies, applying your skills to bridge data engineering and analytics, and making an impact on projects in sports technology and business strategy.
Key Responsibilities:
·      Participate in the design, development, and maintenance of Rapsodo’s data systems, using tools like Google BigQuery, Kafka, and GCP Pub/Sub.
·      Support data ingestion, processing, and ETL (Extract, Transform, Load) processes to deliver reliable data for analytics teams.
·      Collaborate with data analysts and scientists to ensure clean, structured data is available for analysis, reporting, and visualization.
·      Assist in the setup and monitoring of real-time data pipelines, learning to identify ways to streamline data ingestion and reduce latency.
·      Work on data collection, cleaning, and preprocessing, preparing data for analytics and helping generate insights for diverse projects.
·      Collaborate with the team to troubleshoot and improve data pipeline and infrastructure efficiency.
·      Create basic reports, dashboards, and visualizations with tools like Tableau or Power BI, conveying valuable insights to business stakeholders.
·      Currently pursuing or recently completed a Bachelor's or Master's degree in Computer Science, Data Engineering, Data Science, or a related field.
·      Knowledge of SQL and familiarity with Google BigQuery or other data warehousing solutions.
·      Exposure to programming languages such as Python or R and experience with data manipulation, cleaning, and preprocessing techniques.
·      Interest in or exposure to data streaming technologies (e.g., Kafka, GCP Pub/Sub) and ETL development.
·      Basic understanding of statistical analysis and hypothesis testing to support data-driven insights.
·      Experience with data visualization tools (e.g., Apache Superset, Tableau, Power BI, or Matplotlib) is a plus.
·      Excellent written and verbal communication skills.
·      Strong analytical skills, attention to detail, and problem-solving abilities.
·      Strong interpersonal skills to collaborate effectively with cross-functional teams and stakeholders.
·      An inquisitive mindset, coupled with a desire for continuous learning and growth.