Rapsodo is a Sports Technology company with offices in the USA, Singapore, Turkey & Japan. We develop sports analytics products that are data-driven, portable and easy-to-use to empower athletes at all skill levels to analyse and improve their performance. From Major League Baseball star pitchers to Golf tour players, athletes use Rapsodo technology to up their game across the world. Trusted by coaches and players from youths to professionals, Rapsodo provides real-time insights for all-time performance.
We are innovative, focused, and rapidly growing. We are continuously looking for team players who will stop at nothing to deliver state-of-the-art solutions as part of Team Rapsodo.
About the role:
As a Data Engineering Intern at Rapsodo, youโll play an essential role in building and optimizing the data infrastructure that powers decision-making and innovation throughout our organization. This role is ideal for students passionate about data engineering, with a foundational knowledge of data analytics to support meaningful insights. Youโll gain hands-on experience with cutting-edge technologies, applying your skills to bridge data engineering and analytics, and making an impact on projects in sports technology and business strategy.
Key Responsibilities:
ยทย ย ย ย ย ย Participate in the design, development, and maintenance of Rapsodoโs data systems, using tools like Google BigQuery, Kafka, and GCP Pub/Sub.
ยทย ย ย ย ย ย Support data ingestion, processing, and ETL (Extract, Transform, Load) processes to deliver reliable data for analytics teams.
ยทย ย ย ย ย ย Collaborate with data analysts and scientists to ensure clean, structured data is available for analysis, reporting, and visualization.
ยทย ย ย ย ย ย Assist in the setup and monitoring of real-time data pipelines, learning to identify ways to streamline data ingestion and reduce latency.
ยทย ย ย ย ย ย Work on data collection, cleaning, and preprocessing, preparing data for analytics and helping generate insights for diverse projects.
ยทย ย ย ย ย ย Collaborate with the team to troubleshoot and improve data pipeline and infrastructure efficiency.
ยทย ย ย ย ย ย Create basic reports, dashboards, and visualizations with tools like Tableau or Power BI, conveying valuable insights to business stakeholders.
ยทย ย ย ย ย ย Currently pursuing or recently completed a Bachelor's or Master's degree in Computer Science, Data Engineering, Data Science, or a related field.
ยทย ย ย ย ย ย Knowledge of SQL and familiarity with Google BigQuery or other data warehousing solutions.
ยทย ย ย ย ย ย Exposure to programming languages such as Python or R and experience with data manipulation, cleaning, and preprocessing techniques.
ยทย ย ย ย ย ย Interest in or exposure to data streaming technologies (e.g., Kafka, GCP Pub/Sub) and ETL development.
ยทย ย ย ย ย ย Basic understanding of statistical analysis and hypothesis testing to support data-driven insights.
ยทย ย ย ย ย ย Experience with data visualization tools (e.g., Apache Superset, Tableau, Power BI, or Matplotlib) is a plus.
ยทย ย ย ย ย ย Excellent written and verbal communication skills.
ยทย ย ย ย ย ย Strong analytical skills, attention to detail, and problem-solving abilities.
ยทย ย ย ย ย ย Strong interpersonal skills to collaborate effectively with cross-functional teams and stakeholders.
ยทย ย ย ย ย ย An inquisitive mindset, coupled with a desire for continuous learning and growth.