Purpose of the role
To build and maintain the systems that collect, store, process, and analyse data, such as data pipelines, data warehouses and data lakes to ensure that all data is accurate, accessible, and secure.
Accountabilities
Analyst Expectations
All colleagues will be expected to demonstrate the Barclays Values of Respect, Integrity, Service, Excellence and Stewardship – our moral compass, helping us do what we believe is right. They will also be expected to demonstrate the Barclays Mindset – to Empower, Challenge and Drive – the operating manual for how we behave.
Join us as a "Senior Software Engineer" at Barclays, where you'll spearhead the evolution of our digital landscape, driving innovation and excellence. You'll harness cutting-edge technology to revolutionize our digital offerings, ensuring unapparelled customer experiences.
To be successful as a "Senior Software Engineer", you should have experience with:
Technical Skills: Expertise in SPARK/Scala development through the full software development lifecycle (SDLC), including design, development, testing, and deployment.
ETL & Data Pipelines: Build and optimize ETL processes and data pipelines for large-scale data processing using Apache Spark, Scala, Hive, and Airflow.
DevOps & CI/CD: Implement and manage DevOps pipelines using tools like Gitlab, SBT, or Maven.
Code Quality & Best Practices: Conduct code reviews, enforce coding standards, and ensure test coverage and documentation.
Stakeholder Collaboration: Work closely with business analysts, product owners, and other stakeholders to help translate requirements into technical solutions.
Architecture & Design: Implement scalable architectures using Scala, Spark, and Hadoop ecosystems.
Essential Skills:
Languages & Frameworks: Scala (core), Spark, Hive, SQL, Snowflakes, AWS.
Tools: Apache Airflow, Gitlab, SBT, Maven.
Cloud & Microservices: Experience with cloud adoption and microservice architecture is mandatory.
Working knowledge of AWS.
Core AWS service - EC2, Lambda, ECS, EKS ( Kubernetes on AWS).
Storage & database -S3, Document DB.
Networking -VPC, Route53, API Gateway, CloudFront.
Data Engineering & processing -> AWS Glue / Athena/ EMR.
Event processing -SNS, SQS, Event Bridge etc.
Security & IAM - Policies, roles, fine-grained access, KMS encryption etc.
DevOps -CloudFormation/CDK/Terraform, CloudWatch etc.
Experience of working in Gen AI & Agentic AI technologies.
Deep understanding of LLMs, SMLs, Embedding models, Classification models etc.
Proficient in Prompting techniques and prompt engineering.
Preferred expertise in vibe-coding.
Understanding of Agentic Architectures.
Hands-on experience of any of AI Agent Frameworks like Autogen, CrewAI etc.
Working experience on building LLM based apps on any cloud platforms like Amazon Bedrock, Google Vertex etc.
Hands-on working experience on LLM orchestration and RAG frameworks like Lang chain, Lang flow, Haystack etc.
Working experience on RAG based workflows and MCP.
Domain Knowledge: Familiarity with Risk, Finance, and Treasury systems is highly desirable.
Location-Pune.
You may be assessed on the key critical skills relevant for success in role, such as risk and controls, change and transformation, business acumen strategic thinking and digital and technology, as well as job-specific technical skills.
Get personalized insights and tailored applications with our AI tools:
Get your exact compatibility score for each job based on your CV and experience
Automatically optimize your CV for each specific job application
Identify missing skills and get actionable improvement recommendations
No credit card required • 100% free to start
Join thousands of professionals getting AI-powered job recommendations tailored to their skills.