POSITION DESCRIPTION
Critical role designing, developing, and maintaining scalable data pipelines, data sets, and systems for the company’s data infrastructure. Will be collaborating with stakeholders to understand data requirements; Develop ETL (Extract, Transform, Load) processes to change raw data into usable formats for analysis, use and cost effectiveness utilizing techniques such as partitioning, indexing, and caching; Design and implement data models to support the business analytical and reporting needs; ensure data integrity, consistency, and performance and identify errors and performance issues; perform data quality checks for accuracy and consistency of data across systems and platforms; Work closely with business analysts/software engineers to understand their data needs and provide technical expertise and support. Document data pipelines, processes, and best practices for knowledge sharing and future reference; Identify and resolve issues to ensure uninterrupted data flow and availability. Relocation/ telecommuting may be required plus travel to various unanticipated client locations within the United States for short- and long-term assignments.
LANGUAGES, SKILLS AND TOOLS
3 or more of the following tools shall be utilized: Oracle, SQL Server, UNIX, Python, AWS, Azure, Git, Apache (Spark, Kafka, Airflow), Google Big Query. SQL, ETL, Informatica, DB2, Snowflake.
EDUCATION
Master’s degree in computer science/computer applications/engineering (Mechanical/ Information Technology/Civil/ Electrical/Electronics). Will accept bachelor’s degree in computer science/computer application(s)/Engineering (Mechanical/ Information Technology/Civil/ Electrical/Electronics) plus five (5) years of progressive experience in related fields in lieu of Master’s Degree. Will accept foreign education equivalent.