Available for opportunities

Senior Integration & QA Automation Engineer

Faris
Khaleel

|

I design and test integrations where reliability is non-negotiable—powering critical systems across APIs, platforms, and large-scale environments.

Abu Dhabi, UAE
Faris Khaleel
Open to roles
Integration · QA · Product
4+ Years Exp.
Scroll

I specialize in complex integrations
where failure is not acceptable.

  • Led API integrations across 17 insurers and 7 government entities — from schema design through production validation, across UAE and KSA
  • Built QA automation from scratch — covering 500+ endpoints and 200+ performance-tested APIs, simulating up to 3,000 concurrent users
  • Delivered systems handling millions of requests per hour across Insurtech, Proptech, and WellnessTech platforms

Integrations fail silently and at the worst times. I build the validation infrastructure that catches failures before production — and the automation that keeps them from coming back.

API Ecosystem Design

Architecting scalable integration schemas that translate partner requirements into production-grade APIs.

Performance Testing

Load & stress testing at scale — millions of requests/hour, realistic traffic patterns, full observability.

AI-Powered Automation

Built AI-powered OCR and documentation tools that cut manual integration onboarding effort significantly.

QA Leadership

From test strategy to production sign-off — establishing QA from zero at startups and enterprises.

Selected Work

Production systems. Real constraints. Business impact.

Insurtech · Shory Insurance Brokers

Multi-Insurer API Integration Platform — UAE & KSA

Each insurer had a completely different API contract — different auth schemes, payload formats, and error models. A failed integration meant a policy could not be issued. I owned the full integration lifecycle: schema design, automated validation, and production sign-off for every partner API.

160+APIs Integrated
17Insurers Live
7Gov. APIs Delivered

The Problem

Shory needed to bring Motor, Health, and Travel insurance online across two markets fast — but each insurer used a completely different API structure: different authentication (API keys, OAuth, mutual TLS), inconsistent field naming, and error codes with no standard. Government integrations in UAE and KSA added regulatory schema requirements on top. Manual onboarding took weeks per partner, was error-prone, and had no repeatable validation process to catch failures before they reached production users.

Integration Architecture

PARTNER API Raw spec · varied format GOVT API Regulatory schema Integration Layer Schema Translation Auth Normalisation Payload Mapping Error Handling Validation Suite Postman · Newman · SoapUI ✓ PASS 200 OK · Schema valid Auth confirmed ✗ FAIL 4xx/5xx · Field missing → Blocked, not shipped PRODUCTION Policies issued live 17 insurers active Integration Guides Partner-facing docs

Built a structured integration pipeline: translated raw partner specs (REST, SOAP, PDF) into normalised internal schemas, then ran every integration through a validation suite covering auth, payload structure, edge cases, and error handling before any partner went live. No integration reached production without passing the full validation suite.

Sample Integration Payload (sanitized)

POST /api/v1/quotes/motor/initiate Internal → Insurer API
{
  "request_id": "req_xK9mP2qL7nR",
  "insurer_code": "INS_[REDACTED]",
  "product": "MOTOR_COMPREHENSIVE",
  "applicant": {
    "id_type": "EMIRATES_ID",
    "dob": "1990-01-01",
    "nationality": "AE"
  },
  "vehicle": {
    "make": "Toyota",
    "model": "Camry",
    "year": 2021,
    "chassis_no": "[REDACTED]"
  }
}

// Validated: auth header, required fields, date format, enum values
// Expected: 200 OK with premium breakdown + quote_id

Business Impact

  • 160+ API integrations across Motor, Health, and Travel — every one validated before production deployment
  • 17 insurers onboarded end-to-end across UAE & KSA, and 7 government entities
  • Reusable validation suites meant each new insurer onboarding re-used existing test coverage rather than starting from scratch
  • Partner-facing integration guides allowed insurers to self-serve through the integration process with minimal back-and-forth
Stack: PostmanNewmanSoapUIPython JavaScriptSQL ServerAzure DevOpsDockerJira
Insurtech · Shory Insurance Brokers

AI Document Intelligence — OCR + LLM Structured Extraction

Insurance applications required verified customer documents — Emirates IDs, labour contracts, visas, and residence permits. Extracting the right fields manually was slow, inconsistent, and a bottleneck at every stage. I built a two-stage pipeline: OCR to extract raw text from document scans, then a locally-hosted LLM via Ollama to analyse the text and return structured JSON back to the backend.

4+Document Types Automated
LocalLLM via Ollama
JSONStructured Output to Backend

The Problem

Every insurance application required verified identity and residency documents — Emirates IDs, labour contracts, visas, and residence permits. Reading these manually to extract fields like name, ID number, expiry date, and employer was a slow, error-prone process that created a bottleneck at every customer onboarding step. A misread field meant application delays or rejected submissions. I built a two-stage automated pipeline to eliminate this: first, an OCR layer to extract raw text from document scans regardless of layout; then a locally-hosted LLM (Ollama) to semantically understand the extracted text and return only the specific structured fields the backend needed — no hallucinations, no ambiguity, deterministic JSON output.

Pipeline Architecture

Emirates ID Labour Contract Visa Residence Permit Stage 1 — OCR Raw text extraction Python · layout-agnostic Raw Text Unstructured · full doc content extracted Stage 2 — LLM Semantic analysis Field identification Ollama · local inference no external API calls Structured JSON "full_name": "..." "id_number": "..." "expiry_date": "..." "nationality": "..." → Backend API

Built a two-stage pipeline in Python. Stage 1 used OCR to extract raw text from scanned documents regardless of layout — handling a variety of document formats, orientations, and print qualities across Emirates IDs, labour contracts, visas, and residence permits. Stage 2 passed that raw text to a locally-hosted LLM running via Ollama, prompted to identify and extract only the specific fields required by the backend (name, ID number, expiry, nationality, employer, etc.) and return them as a deterministic structured JSON object. No external API calls — the LLM ran locally, keeping all document data on-premise.

Sample Structured Output

OUT Ollama LLM → Backend Structured JSON Response
{
  "document_type": "emirates_id",
  "full_name": "[REDACTED]",
  "id_number": "784-[REDACTED]",
  "nationality": "[REDACTED]",
  "date_of_birth": "[REDACTED]",
  "expiry_date": "[REDACTED]",
  "confidence": "high"
}

// LLM prompted with field schema — returns only required fields
// Ollama running locally — zero external data exposure

Business Impact

  • 4 document types fully automated — Emirates IDs, labour contracts, visas, and residence permits
  • Eliminated manual document reading from the onboarding workflow — zero human transcription needed for standard documents
  • Two-stage design (OCR → LLM) handled layout variation and print quality issues that single-step extraction could not
  • Local Ollama inference kept all customer document data on-premise — no third-party API exposure for sensitive PII
Stack: PythonOllamaOCRREST APIs SQL ServerAzure DevOps
Proptech · Skyloov

End-to-End QA Infrastructure — Built from Zero

Skyloov had no QA processes, no test automation, and no performance baselines — while running a high-traffic platform with 120K+ property listings. As the sole QA engineer, I built the entire testing infrastructure from scratch and established the first-ever performance benchmarks for the platform.

500+Endpoints Covered
1,000Concurrent Users Tested
120KListings at Stake

The Problem

Skyloov was shipping high-risk releases with no test coverage and no performance data, across a platform handling 120K+ property listings. Bugs were reaching production users regularly. There was no definition of what "good performance" looked like, and no way to catch regressions between releases. I was hired as the only QA engineer in a ~12-person engineering team to change this — from day one.

Testing Pipeline Architecture

FUNCTIONAL REGRESSION Dev Commit GitHub push Jenkins Trigger CI pipeline start Newman Run 500+ endpoints tested Test Report Pass/Fail per endpoint Release Gate Block or approve deploy PERFORMANCE TESTING Gatling Script 1,000 virtual users Load Execution 200+ API endpoints Metrics Collected p95 latency · TPS · Errors Grafana Real-time dashboards Baseline Report Per-release benchmarks Also validated: ElasticSearch search relevance · Kafka async workflows · AI chatbot flows

Built two parallel testing tracks that ran every release cycle. Functional track: Postman collections covering 500+ backend endpoints, executed automatically via Newman in Jenkins CI. Performance track: Gatling simulating 1,000 concurrent users across 200+ APIs with p95 latency, TPS, and error rate metrics exported to Grafana dashboards. For the first time, the team had a performance baseline they could compare against after every deployment.

Business Impact

  • 500+ endpoint regression suite running on every deployment — regressions caught before reaching 120K+ property listings users
  • Platform launched successfully across 3 apps (web, iOS, Android) with established pass/fail gates at every release
  • First-ever performance baselines defined: p95 latency, max throughput, error rate thresholds enforced per release
  • AI-powered property search chatbot fully validated across search, support, and lead generation flows before launch
  • ElasticSearch relevance and Kafka async event processing validated under real workload conditions
Stack: PostmanNewmanGatlingGrafana JenkinsElasticSearchKafkaPostgreSQL MongoDBDockerCharles Proxy
Performance · Aspire IT Services

Peak-Traffic Performance Testing for Microservices Backend

A microservices-based backend needed to withstand high-traffic peak events, but the existing performance testing setup lacked baselines and visibility into system bottlenecks. I extended and matured the platform into a comprehensive performance testing solution.

3,000Concurrent VUsers
M+Requests per Hour
100sof Endpoints Tested

The Problem

The team anticipated high-traffic periods, but lacked confidence in how the system would perform under load. Existing performance testing was limited, with no clear targets or sufficient observability into system behavior under stress. Microservices were tightly coupled, increasing the risk of cascading failures. I was brought in to evolve and scale the performance testing program, strengthening coverage, defining benchmarks, and improving visibility ahead of peak periods.

Performance Testing Architecture

Gatling Scripts Scala · Ramp / Spike / Sustained load scenarios 3,000 VUs Concurrent virtual users M+ requests / hour Microservices Auth Service User Service Product API Search Service Order / Checkout Notification Service + more Grafana CPU · Memory · p95 Latency Throughput · Error Rate New Relic APM Distributed traces Service-level bottlenecks Bottlenecks Found Baselines established Fixes prioritised Auto-Published Reports GitHub Pages · HTML

Wrote Gatling scripts in Scala covering hundreds of endpoints across all microservices. Designed three scenario types: ramp-up (gradual user growth), sustained load (steady-state at target TPS), and spike (sudden 3,000 VU burst). Monitored CPU, memory, p95 latency, throughput, and error rates simultaneously using Grafana dashboards and New Relic APM traces. Automated test report generation and publishing to GitHub Pages — so the team had full results immediately after every run without any manual steps.

Business Impact

  • 3,000 concurrent virtual users generating millions of API requests per hour — revealing real bottlenecks before any traffic event hit production
  • Multiple microservice-level bottlenecks identified and fixed before peak periods — preventing cascading failures in production
  • System performance baselines established for the first time: latency targets, max throughput, memory ceilings per service
  • Automated report pipeline meant the entire team had access to test results instantly — no manual reporting overhead
Stack: Gatling (Scala)GrafanaNew Relic AWSDockerJenkinsCypressGoogle Lighthouse
WellnessTech · Aspire IT Services

Automated ETL Validation & Data Quality Framework — WeightWatchers

A global wellness platform's analytics pipeline spanned Snowflake, BigQuery, and Oracle — but had no automated validation to catch schema drift, data reconciliation failures, or duplicate records before they reached dashboards and reporting. I built a Python-based data quality framework to automate these checks across all environments.

3DB Environments Covered
100%Automated ETL Checks
6BI Migrations Validated

The Problem

WeightWatchers ran analytics and reporting across a multi-cloud data stack — Snowflake, BigQuery, and Oracle — feeding business dashboards and user-facing reporting features. ETL pipelines moved data between these environments, but there were no automated checks to verify correctness after each run. Schema drift, duplicate records, and reconciliation mismatches were only discovered reactively — often after bad data had already surfaced in reports. The team was also mid-migration from DSS Reporting to Looker Studio, with no validation framework to confirm the migrated dashboards matched source truth. I was brought in to build the validation infrastructure from scratch.

Data Validation Pipeline Architecture

Snowflake Analytics warehouse BigQuery GCP data platform Oracle DB Operational source ETL Pipeline Extract → Transform → Load Python automated Validation Engine Schema checks Row reconciliation Duplicate detection Python · SQL ✓ Data Valid Pipeline output confirmed ✗ Issues Found Drift / dupes → flagged Looker Studio Migrated from DSS Dashboards validated

Wrote Python-based ETL validation scripts that ran automated checks across Snowflake, BigQuery, and Oracle after every pipeline execution: schema drift detection, row count reconciliation between source and destination, duplicate record identification, and field-level data type checks. For the BI migration from DSS Reporting to Looker Studio, validated that every dashboard metric matched its source dataset — ensuring the migration produced no regressions in reported figures.

Business Impact

  • 3 database environments covered — Snowflake, BigQuery, and Oracle — under a single automated validation framework
  • Shifted ETL quality assurance from reactive (catch issues in reports) to proactive (catch failures at the pipeline boundary)
  • Schema drift, duplicate records, and reconciliation mismatches caught before reaching analytics dashboards or end-user reports
  • DSS → Looker Studio BI migration validated end-to-end — confirmed dashboard metrics matched source data after full migration
Stack: PythonSQLBigQuerySnowflake Oracle DBMongoDBLooker StudioGit

Measurable Engineering Impact

160+
APIs Integrated & Validated
Motor, Health & Travel — every one validated before go-live
17
Insurance Partners Onboarded
End-to-end across UAE & KSA
500+
Endpoints Under Automated Test
Postman & Newman regression on every deployment
3,000+
Concurrent Virtual Users Simulated
Generating millions of requests/hour — Gatling stress tests
120K
Property Listings Tested
High-traffic platform — search, AI chatbot, mobile & web
7
Government Entities Integrated
Regulatory entities across UAE & KSA

Where I've Shipped

Integration Engineer

Shory Insurance Brokers (First.tech Group)
Nov 2024 — Present Abu Dhabi, UAE
Insurtech
  • Onboarded 17 insurance companies end-to-end — Motor, Health, and Travel across UAE & KSA
  • Designed and validated 160+ API integrations, and integrating with 7 government entities (UAE & KSA regulatory entities)
  • Built AI-powered OCR system and documentation platform, accelerating partner onboarding
  • Implemented automation workflows for API testing, improving reliability and engineering velocity
  • Led technical discussions on API design, failure investigation, and root cause analysis
Core Postman SoapUI Newman Python JavaScript
Supporting Azure DevOps Docker SQL Jira Figma

Quality Assurance Engineer

Skyloov Property Portal
Jul 2023 — Oct 2024 Dubai, UAE
Proptech
  • Built API regression automation covering 500+ backend endpoints as sole QA engineer in a 12-person startup
  • Designed performance tests for 200+ APIs simulating 1,000 concurrent users — established first-ever performance baselines
  • Supported launch of platform managing ~120K property listings across 3 apps (web, iOS, Android)
  • Validated AI-powered chatbot for property search and Kafka async workflows with ElasticSearch
Core Postman Gatling Grafana Newman
Supporting ElasticSearch Kafka PostgreSQL MongoDB AWS Docker Jenkins

Data Quality Engineer

Aspire IT Services — WeightWatchers Project
Jan 2023 — Feb 2024 Amman, Jordan
WellnessTech
  • Automated ETL validation checks using Python across Snowflake, BigQuery, and Oracle environments
  • Validated analytics and reporting datasets for a global platform — schema checks, reconciliation, duplicate detection
  • Tested and validated BI dashboard migration from DSS Reporting to Looker Studio
Core BigQuery Snowflake Python SQL
Supporting Oracle DB MongoDB Looker Studio Git

Performance Engineer

Aspire IT Services
Oct 2022 — Jan 2023 Amman, Jordan
Performance
  • Simulated 3,000 concurrent virtual users, generating millions of API requests per hour across microservices
  • Identified performance bottlenecks using Grafana and New Relic; collaborated with engineering on remediation
  • Automated Gatling report publishing via GitHub Pages for team-wide post-test visibility
Core Gatling (Scala) Grafana New Relic
Supporting AWS Docker Jenkins Cypress Google Lighthouse

QA Engineer

Aspire IT Services
Jul 2022 — Sep 2022 Amman, Jordan
QA
  • Functional and API testing for mobile apps and backend services
  • Built Postman scripts for API validation; sprint testing in Agile environment
Core Postman Jira Charles Proxy
Supporting Swagger PostgreSQL Confluence

Skills & Expertise

Core Expertise
Supporting Tools

Databases & Data

PostgreSQL BigQuery Snowflake MongoDB Oracle DB ElasticSearch MS SQL Server

Infrastructure & CI/CD

Docker Jenkins AWS Azure DevOps Kafka Git / GitHub Linux

Platforms & Tools

Jira (Admin) Figma Looker Studio TestFairy TestFlight Confluence

Bachelor's in Computer Science

University of Jordan

Feb 2018 — Jul 2022

Toastmasters International

Member — 3+ years across Amman, Dubai & Abu Dhabi

Awarded Best Prepared Speaker — 195th Club Session

Ready to Work Together?

Open to senior integration, QA automation, and API engineering roles across the UAE and beyond.
Reach out — I respond fast.