KEERTHI JAKKA
⋄ +1 (908) 808-9662 ⋄ USA ⋄ Gmail: jsri8991@[Link]⋄
PROFESSIONAL SUMMARY
● Worked exclusively with a financial client, focusing on optimizing data workflows, ensuring compliance,
and delivering high-quality data solutions tailored to the industry’s regulatory and business needs.
● Designed and implemented scalable ETL/ELT pipelines to extract, transform, and load data into Teradata
from cloud-based sources like Azure Blob Storage and Snowflake, ensuring seamless integration and
transformation.
● Optimized complex Teradata SQL queries for improved performance, leveraging partitioning and
indexing capabilities to enhance query execution speed and reduce resource utilization.
● Built robust data marts by implementing ETL processes with Teradata, integrating structured and
unstructured data sources while ensuring data quality and consistency.
● Automated data ingestion and transformation processes using Teradata SQL Assistant, reducing manual
intervention, improving operational efficiency, and enabling real-time data updates.
● Designed and implemented data pipelines using Azure Data Factory and Azure Databricks to integrate
Teradata with other data sources, enabling seamless data migration and advanced processing.
● Utilized Teradata's analytical capabilities alongside big data tools like Spark and Kafka, delivering high-
performance analytics solutions for large-scale datasets.
● Designed and optimized graph-based data models using GSQL, enabling efficient querying and
visualization of complex relationships across large datasets.
● Developed advanced GSQL queries for tasks such as shortest path analysis, community detection, and
subgraph extraction to support business intelligence initiatives.
● Integrated GSQL into ETL/ELT workflows, ensuring compatibility with tools like Snowflake, Teradata, and
Azure Data Factory to enable seamless graph data processing.
● Enhanced graph query performance by tuning GSQL scripts and leveraging parallel processing capabilities,
reducing execution time and resource usage.
TECHNICAL SKILLS
Programming Languages HTML, CSS, Java 8, React , JavaScript
Terradata Expertise Terradata BTEQ, Fast Load, Multi Load, TPump, TPT, Query Optimization, Partitioning,
ETL Pipelines
Data Warehousing & ETL Talend, Snowflake
Scripting Languages Shell Scripting
Cloud Technologies AWS(EC2,EMR,Redshift), GCP
Hadoop Distributions Cloudera, Hortonworks
Data Visualization Tools Tableau, Power BI
Operating Systems Linux, Windows,Unix,Ubuntu
Version Control Github
PROFESSIONAL EXPERIENCE
Client: Ford Credit Aug 2023 – Present
Role: Software Developer
Responsibilities:
Proficient in developing robust, scalable applications using Spring Boot, Java 8+, Apollo GraphQL, and
Postman for API integration, with expertise in database optimization using SQL Server and Teradata.
Designed and monitored microservices with Swagger UI and Springboard Console, and implemented
CI/CD pipelines using Jenkins, Gradle, and GCP for streamlined deployments and enhanced system
reliability.
Led teams in adopting Agile (SCRUM) methodology, managing end-to-end SDLC phases, and ensuring
timely delivery and alignment with business goals.
Skilled in deploying and managing cloud-based applications on AWS and GCP, and leveraging monitoring
tools like Dynatrace and Splunk to ensure high system performance and availability.
Developed and optimized ETL pipelines for loading, transforming, and analyzing large datasets in
Teradata using BTEQ, FastLoad, and MultiLoad, ensuring data consistency and high performance.
Integrated structured and unstructured data into Teradata data warehouses, utilizing partitioning and
indexing to enhance query execution and storage efficiency.
Automated data ingestion and transformation processes in Teradata using SQL scripts and workflows,
reducing manual intervention and improving operational efficiency.
Worked extensively with Teradata TPT to enable seamless data migration between cloud sources and on-
premises systems, supporting scalable analytics.
Integrated GSQL-based graph databases with microservices and APIs, ensuring seamless
communication between application layers and efficient data exchange.
Optimized GSQL queries for performance by leveraging parallel processing, indexing, and partitioning
techniques to handle high-volume graph data efficiently.
Automated graph data ingestion workflows using GSQL scripts and orchestration tools, reducing manual
interventions and improving data availability.
Client: University of Alabama in Huntsville Aug 2021 – April 2023
Role: Site Reliability Engineer
Responsibilities:
Taught students how to deploy and manage cloud resources, such as virtual machines, storage, databases,
and networking components, using tools like AWS EC2, S3, and RDS, with a focus on integrating cloud
data into Teradata environments for analytics.
Educated students on serverless computing concepts, guiding them in using services like AWS Lambda
and Azure Functions to build event-driven, scalable applications with seamless data migration to Teradata.
Provided hands-on training on cloud security best practices, including utilizing tools like AWS Identity
and Access Management (IAM) for access control and permissions management, ensuring compliance
with secure data handling practices for Teradata projects.
Delivered engaging lectures on Python fundamentals, emphasizing data processing and ETL pipeline
development for Teradata-centric architectures.
Ensured the smooth operation of computer science labs, offering technical support for students working on
big data projects involving Teradata, SQL, and Hadoop ecosystem components.
Guided students in designing ETL pipelines using Teradata BTEQ, FastLoad, and MultiLoad, and
demonstrated how to optimize queries for large-scale datasets in Teradata environments.
Introduced students to advanced data warehousing concepts, including Teradata's partitioning and indexing
capabilities, for building high-performance and scalable data models.
Demonstrated the use of GSQL in real-time analytics scenarios, such as fraud detection, recommendation
systems, and network analysis, showcasing the value of graph databases in modern data-driven
applications.
Educated students on schema design in graph databases, emphasizing the importance of node and edge
definitions, attributes, and relationships for accurate data modeling.
Introduced visualization tools integrated with GSQL, helping students create interactive dashboards to
explore and analyze graph-based metrics like centrality, clustering, and hierarchical structures.
Explored hybrid data workflows by combining relational data (Teradata) and graph data (GSQL), enabling
students to understand and apply both paradigms for comprehensive data insights.
Mentored students in using GSQL APIs to integrate graph database functionalities into web applications
and microservices, emphasizing the practical implementation of graph analytics in software development
Client: University of Alabama in Huntsville May 2021 – Aug 2021
Role: Software Developer
Responsibilities:
● Aided in research for NASA’s South/Southeast Asia Research Initiative (SARI), contributing to projects
focused on environmental and climate analysis.
● Calculated and visualized Normalized Difference Vegetation Index (NDVI) values for various geographical
datasets using satellite imagery from Landsat 5, Landsat 7, Landsat 8, MODIS, and Sentinel 2, enabling
detailed vegetation health assessments.
● Processed large-scale satellite data using remote sensing tools and programming languages such as Python
and R, automating data extraction and NDVI computation workflows.
● Conducted data cleaning, transformation, and validation to ensure accuracy and reliability in NDVI analysis
for diverse geospatial regions.
● Utilized GIS software such as QGIS or ArcGIS to map NDVI values and overlay them with other
environmental layers for comprehensive spatial analysis.
● Collaborated with a multidisciplinary team to integrate satellite-based observations into research studies,
contributing to actionable insights for regional ecological monitoring and agricultural planning.
● Explored trends in vegetation patterns over time by performing temporal analyses using datasets from
multiple satellites, aiding in climate impact assessments.
● Documented methodologies and findings in detailed reports, contributing to the creation of reproducible
workflows and datasets for future research initiatives.
Client: Sunseaz Technologies PVT ltd Jan 2018 – Dec 2020
Role: Software Developer
Responsibilities:
● Executed big data analytics and predictive analytics initiatives, leveraging Python and Spark for
advanced data processing and analysis.
● Developed and optimized Spark jobs using Python for data validation, cleansing, transformation, and
aggregation, ensuring efficient data processing.
● Implemented real-time data streaming solutions using Apache Kafka, Spark Streaming, and AWS
Kinesis, enabling fast and reliable data ingestion.
● Developed and maintained ETL processes using SQL, DataStage, and Python, ensuring efficient data
extraction, transformation, and loading.
● Configured and scheduled Talend jobs for data integration, optimizing for performance and scalability in
large data environments.
● Automated data ingestion processes using Python and Scala, integrating data from various sources such as
APIs, AWS S3, and Snowflake.
● Implemented SCD mechanisms for managing changing dimensions in data warehouses, ensuring
accurate historical reporting.
● Incorporated GSQL into DevOps practices, demonstrating the integration of graph database deployments
and updates into CI/CD pipelines for seamless operational workflows.
● Enabled advanced graph analytics by implementing GSQL-based algorithms such as PageRank, shortest
path, and connected components, providing actionable insights for business processes.
● Deployed graph databases with GSQL in cloud environments like AWS and GCP, leveraging
containerization and CI/CD pipelines for streamlined deployments and updates.
● Integrated GSQL with front-end services via APIs, ensuring smooth interaction between graph databases
and user-facing applications for seamless data visualization and insights.
CERTIFICATION
Apollo Certified Graph Developer - Associate
Apollo Certified Graph Developer - Professional