HomeFresher JobsSnowflake Off Campus Drive 2026 | Data Engineer Intern | Pune —...

Snowflake Off Campus Drive 2026 | Data Engineer Intern | Pune — Apply Now

Live Updates Join WhatsApp for Daily Job Alerts
JOIN NOW

Snowflake Off Campus Drive 2026 is now open for engineering graduates looking to build a strong career in data engineering. Snowflake, a global leader in cloud data warehousing, is hiring for the role of Data Engineer Intern at its Pune office in Pune. This is an excellent opportunity to work on large-scale data systems and gain hands-on experience with modern data infrastructure in a real production environment.

Job Overview

  • Company: Snowflake
  • Role: Data Engineer Intern
  • Location: Pune, Maharashtra, India
  • Eligibility: B.Tech / B.E. / M.Tech (CS, IT, or related fields)
  • Batch: 2024 / 2025 / 2026
  • Experience: Freshers / Internship
  • Salary / Stipend: Competitive (Not Disclosed)
  • Work Mode: Work From Office
  • Apply Mode: Online
  • Last Date: Apply as soon as possible

About Snowflake

Snowflake is a cloud-native data platform company that has revolutionized how organizations store, process, and analyze data. Founded in 2012 and headquartered in the United States, Snowflake offers a unified Data Cloud platform that operates seamlessly across major cloud providers like AWS, Microsoft Azure, and Google Cloud.

The company serves thousands of global enterprises and is known for its strong engineering culture, innovation-driven approach, and rapid growth. Its India operations, especially in Pune, play a key role in building scalable and high-performance data solutions.

Role Overview – Data Engineer Intern

As a Data Engineer Intern, you will work on building and managing data pipelines that support Snowflake’s internal analytics systems. You will collaborate with multiple teams including engineering, security, and business stakeholders to ensure data reliability and efficiency.

This role provides exposure to production-level data systems and hands-on experience with real-world data engineering challenges.

Key Responsibilities

  • Data Pipeline Development: Build and maintain scalable data pipelines using Python and Apache Airflow
  • Data Integrity Management: Ensure data accuracy, reliability, and uptime across systems
  • Ingestion Frameworks: Design and implement scalable data ingestion systems
  • Cross-Team Collaboration: Work with engineering, security, and business teams
  • API Integration: Consume REST APIs and integrate external data sources
  • Cloud Engineering: Work with AWS, Azure, or GCP environments
  • Knowledge Sharing: Contribute to documentation and team learning

Eligibility Criteria

Educational Qualification

  • B.Tech / B.E. / M.Tech in Computer Science, IT, or related fields

Batch Eligibility

  • 2024, 2025, 2026 graduates

Required Skills

  • Strong SQL and database concepts
  • Python programming for data workflows
  • Understanding of REST APIs
  • Experience with Apache Airflow

Cloud Knowledge

  • Basic experience in AWS, Azure, or GCP

Preferred Skills

  • Experience with ETL/ELT pipelines
  • M.Tech or advanced degree (optional advantage)

Soft Skills

  • Strong communication
  • Team collaboration ability

Selection Process

The hiring process consists of three main stages:

1. Resume Screening

Applications are evaluated based on technical skills, academic background, and project experience.

2. Technical Interview

Focus areas include:

  • SQL queries
  • Python scripting
  • Data pipeline design
  • Cloud fundamentals

Candidates may also discuss projects or solve practical problems.

3. Final Interview

A discussion with the hiring manager to assess:

  • Technical depth
  • Problem-solving approach
  • Communication and teamwork

Salary and Benefits

Snowflake offers highly competitive compensation for interns. While the exact stipend is not disclosed, selected candidates can expect:

  • Attractive stipend
  • Mentorship from experienced engineers
  • Hands-on work on real data systems
  • Exposure to Snowflake Data Cloud
  • PPO (Pre-Placement Offer) opportunities
  • Flexible and collaborative work environment

How to Apply

Follow these steps to apply:

  1. Visit the official Snowflake careers page
  2. Create or log in to your account
  3. Fill in personal and academic details
  4. Upload your updated resume (PDF)
  5. Submit the application

Make sure your resume highlights:

  • Python projects
  • SQL skills
  • Data pipeline or ETL experience
  • Cloud platform exposure

Application Link

Apply here:
https://careers.snowflake.com/us/en/job/SNCOUS97879D11040B471CAB908A9497E6D01EEXTERNALENUS8904ED8575D648CFB01B7941677FDE4A/

RELATED ARTICLES

LEAVE A REPLY

Please enter your comment!
Please enter your name here

- Advertisment -
Updates Sending Now
Join Our WhatsApp Job Alerts

Be the first to apply! We post 10+ new fresher and experience job openings every single day.

Tap to Join Now
Join 12,450+ job seekers today

Most Popular