What are the responsibilities and job description for the Internship – Data Engineering & Analytics position at General Genomics Inc.?
General Genomics, Inc. (GGI) is an AI / ML firm headquartered in Oklahoma City, OK, with our wholly owned subsidiary General Genomics AI located in Fort Worth, TX. Our mission is to address bioscience, bioinformatics, and national security challenges through innovative technology, enabling clients to analyze and interpret large structured and unstructured data sets efficiently. We specialize in developing customized algorithms that support critical decision-making for various organizational missions. GGI has been recognized as a Small Business Administration AI Accelerator (AIA) in partnership with Oklahoma City's Metro Technology Center (MTC).
Internship Opportunity – Data Engineering & Analytics
Location : Fort Worth, TX (Medical District, Downtown Fort Worth)
Company : General Genomics, Inc. (GGI)
Internship Duration : 12 months (Late May / Early June 2025 – End of May 2026), goal is to onboard the new intern in the middle of May 2025.
Work Schedule : Hybrid (3 days in-office per week during summer, minimum of 2 half-days or one full day per week in-office during the school year)
Hours : Maximum of 20 hours per week during the academic year, not to exceed 25 hours per week in the summer, depending on client workload.
Compensation : Paid internship, commensurate with the candidate's qualifications and experience. No benefits are provided with this role.
Internship Role
As an Intern - Data Engineering & Analytics, you will engage in hands-on projects focused on data processing, analysis, and engineering solutions. You will be an integral part of the data team, collaborating on real-world challenges that impact healthcare, bioinformatics, and national security sectors.
Responsibilities
- Assist in data engineering tasks, including ingestion, transformation, and structuring of large data sets.
- Apply Python and SQL for data analysis and automation.
- Work with data processing frameworks such as Apache Spark / pyspark.
- Implement and optimize workflows using Airflow and dbt.
- Explore and manipulate multi-modal datasets.
- Contribute to problem-solving initiatives requiring innovative, non-standard solutions.
- Participate in team discussions, providing insights and analytical support.
Qualifications
Eligibility Requirements
Additional Details
Application Process
Interested candidates should submit their resume, a cover letter, and unofficial transcripts demonstrating their academic standing.
PI263882699