(GCP) Data Integration Engineer
Company: KANINI
Location: Nashville
Posted on: February 28, 2026
|
|
|
Job Description:
Job Description Job Summary: We are looking for a
detail-oriented and technically skilled GCP Data Integration
Engineer to design, develop, and manage robust data integration
solutions. The ideal candidate will have hands-on experience in
integrating data across disparate systems, building ETL/ELT
pipelines, and ensuring the accuracy, quality, and consistency of
enterprise data. You will play a key role in enabling seamless data
flow between systems to support business intelligence, analytics,
and operational needs. Key Responsibilities: Integration Design &
Development Design and implement data integration workflows between
internal and external systems, including APIs, databases, SaaS
applications, and cloud platforms. Develop and maintain scalable
ETL/ELT pipelines for structured and unstructured data using tools
like Informatica, Talend, SSIS, Apache NiFi, or custom Python/SQL
scripts. Build and manage real-time and batch data pipelines
leveraging technologies like Kafka, Spark Streaming,. Data Quality
& Governance Ensure high data quality, accuracy, and consistency
during data ingestion and transformation. Implement data
validation, cleansing, deduplication, and monitoring mechanisms.
Contribute to metadata management, data lineage, and data catalog
initiatives. Collaboration & Troubleshooting Collaborate with data
engineers, business analysts, data scientists, and application
teams to understand integration needs and deliver effective
solutions. Troubleshoot and resolve data integration and pipeline
issues in a timely manner. Provide documentation and knowledge
transfer for developed solutions. Platform & Infrastructure Support
Support data movement across hybrid environments (on-prem, cloud,
third-party systems). Work with DevOps or platform teams to ensure
scalability, security, and performance of data integration
infrastructure. Qualifications: Bachelor’s degree in Computer
Science, Information Systems, Engineering, or a related field. 4–8
years of experience in data integration, data engineering, with
strong ETL and SQL. Strong experience with integration tools such
as Informatica, Talend, MuleSoft, SSIS, or Boomi. Proficient in
SQL, Python, and scripting for data manipulation and automation.
Experience with cloud data platforms (GCP) and services such as
Google Cloud Dataflow. Familiarity with REST/SOAP APIs, JSON, XML ,
and flat file integrations. Preferred Skills: Experience with
message queues or data streaming platforms (Kafka, RabbitMQ,
Kinesis). Understanding of data warehousing concepts and tools
(e.g., Snowflake, Redshift, BigQuery). Knowledge of data security,
privacy, and compliance best practices (HIPAA, GDPR, etc.). Prior
experience in industries like healthcare, fintech, or e-commerce is
a plus. Soft Skills: Strong problem-solving and debugging skills.
Excellent communication and collaboration abilities. Ability to
manage multiple priorities and deliver in a fast-paced environment.
Attention to detail and a commitment to delivering high-quality
work.
Keywords: KANINI, Longview , (GCP) Data Integration Engineer, IT / Software / Systems , Nashville, Texas