Application Developer

Noida, Uttar Pradesh, India
Nov 21, 2024
Nov 12, 2025
Onsite
Full-Time
3 Years
Job Description

As an Application Developer at Accenture, you will design, build, and configure data-driven applications to meet the business needs of clients using Apache Spark. This is a dynamic role where you will collaborate with cross-functional teams to develop and test scalable, high-performance applications that integrate with business processes. You’ll work with cutting-edge Big Data technologies and contribute to impactful data solutions. Your experience with Apache Spark, Scala, Java, and SQL will be essential in driving successful project delivery.

Key Responsibilities

  • Design, build, and configure applications using Apache Spark to meet the unique requirements of business processes.
  • Write and maintain high-quality code that is scalable, efficient, and optimized for performance.
  • Develop and test data processing applications to ensure they handle large datasets and work efficiently in a distributed environment.
  • Work closely with business analysts, project managers, and other developers to ensure project alignment with business goals and technical specifications.
  • Participate in design and review sessions to contribute to the application architecture and design decisions.
  • Identify and resolve technical issues in applications, using Spark-related tools and frameworks to debug and optimize performance.
  • Provide timely solutions to issues faced during development, testing, and production deployment.
  • Keep up to date with the latest developments in Apache Spark and related Big Data technologies such as Hadoop and Hive.
  • Introduce innovative solutions and techniques that enhance the functionality and performance of applications.
  • Work in Agile environments, collaborating with teams to deliver on sprint goals and project timelines.
  • Participate in daily stand-ups, sprint planning, and retrospectives to ensure timely and high-quality delivery.
  • Ensure the scalability and high-performance of applications by optimizing code, data pipelines, and query processing.
  • Ensure that applications can handle large datasets effectively and deliver high-quality results under varying loads.

Required Qualifications

  • Minimum of 3 years of hands-on experience with Apache Spark in building and deploying distributed data processing applications.
  • Experience with Scala or Java programming languages.
  • Strong understanding of SQL and experience with NoSQL databases.
  • Experience in implementing Big Data solutions using Hadoop, Hive, and other Spark-related technologies.
  • Proven ability to work in an Agile development environment with teams to meet deadlines and deliver impactful solutions.

Technical Skills

  • Must-have skills. Proficiency in Apache Spark.
  • Strong experience with Hadoop, Hive, and related Big Data technologies.
  • Proficient in Scala or Java for data processing and application development.
  • Solid understanding of distributed computing principles and cloud computing (e.g., AWS, GCP).
  • Strong proficiency in SQL and working knowledge of NoSQL databases (e.g., MongoDB, Cassandra).
  • Experience with data pipeline architecture and building scalable systems for data processing.

Soft Skills

  • Excellent problem-solving and troubleshooting skills, particularly in large-scale data environments.
  • Strong communication skills to effectively collaborate with business and technical teams.
  • Ability to take ownership of tasks and ensure high-quality project delivery.
  • Proactive attitude toward learning and applying new technologies.

Educational Qualifications

  • Minimum 15 years of full-time education (e.g., a 10+2+3 or 10+2+4 pattern of education, where applicable).
  • A Bachelor’s degree in Computer Science, Information Technology, or a related technical field is required.
  • A strong academic background with a focus on programming, algorithms, data structures, and distributed computing is essential.

Preferred Qualifications

  • Experience working with cloud platforms (e.g., AWS, Azure, Google Cloud) for data processing and deployment.
  • Knowledge of data security and privacy regulations as they apply to big data technologies.
  • Certification or training in Apache Spark or related technologies would be a plus.
  • Familiarity with DevOps tools and practices for continuous integration/continuous delivery (CI/CD).

Why Accenture?
At Accenture, we believe in fostering innovation, delivering world-class solutions, and creating opportunities for personal and professional growth. As an Application Developer, you will work with leading technologies like Apache Spark and collaborate with top industry professionals to create high-impact data solutions. With a culture that encourages continuous learning and collaboration, you’ll be empowered to grow and make a meaningful contribution to some of the most exciting projects in the tech industry.

How to Apply. When applying for the Application Developer position at Accenture, consider the following tips.

  1. Highlight Spark Expertise. Showcase your experience with Apache Spark, detailing any projects where you developed scalable data processing applications using Spark. Mention how you’ve utilized Spark's RDDs, DataFrames, and SparkSQL for large-scale data transformations.
  2. Showcase Big Data Knowledge. Emphasize your experience with Hadoop, Hive, and other Big Data tools. If you’ve worked on integrating these technologies with Spark, provide examples of how you’ve managed large datasets efficiently.
  3. Programming Skills. Provide specific examples of your work with Scala or Java in data-intensive applications. Mention any frameworks or tools you’ve used to enhance the development and performance of applications.
  4. Agile Experience. Demonstrate your familiarity with Agile practices. Mention specific projects where you contributed to sprints, collaborated with cross-functional teams, and delivered software in a timely manner.
  5. Focus on Problem Solving and Performance. Share examples of how you’ve optimized applications for performance, whether by tuning Spark configurations, optimizing data pipelines, or ensuring efficient resource utilization in a distributed environment.
Related Jobs