

Client
A leading Australian bank
Goal
Streamline the payment transaction lifecycle to handle increasing volume and complexity
Tools and Technologies
Jenkins, Kubernetes, Spring, Oracle, PostgresSQL, AWS, Docker, Kafka, Java
Business Challenge
Multiple banking channels - mobile, internet and branch - initiate various payment requests (20+ types such as ACH, mandate and book transfer) that need to be processed. Depending on the payment type, the channel was required to invoke one or more services in a specific order as per the associated business rules.
Changes to these payment workflows stemming from introduction of new payment types or revisions of business rules introduced complex and repeated changes to the bank’s systems, hindering scalability.

Solution
- To support lifecycle management of various payment transactions, including defining different payment workflows, we designed an event-driven architecture comprised of 40+ microservices (e.g., limit, eligibility, and fraud checks, etc.) supported by a Kafka message queuing system
- The architecture involved building an orchestration engine (landing service), acting as a front controller for all payment workflow requests from the various banking channels, such as mobile, internet and branch
- The landing service in turn invokes the corresponding service (limit, eligibility, etc.) based on the payment type and business rules associated with it
- Data flow between these microservices (resulting from further invocations) and other downstream systems is facilitated asynchronously with the help of a distributed messaging system (Kafka)
- Using Jenkins, we built a CI/CD pipeline to streamline the workflow by automatically building, testing and deploying code changes as they are committed

Outcomes
- Significantly eased the management of payment workflows, including those related to the addition of new payment types (resulting from an acquisition)
- Enabled systems to scale without introducing complex changes at the channels
- Improved reporting, resulting from faster access to data through dedicated microservices

Our experts can help you find the right solutions to meet your needs.
Modernized Payments Hub Improves UX and Compliance



Client
U.S. operations of a leading Japanese bank
Goal
Modernize payments architecture to streamline processing and improve client experience
Tools and Technologies
Jenkins, Kafka, Spring, Oracle, JBoss, React, Elastic Search, Java, Node.js
Business Challenge
The evolving payments landscape, with the introduction of ISO 20022 and the dynamic nature of the regulatory environment, necessitated advancement in the bank’s payment processing capabilities.
The lack of a modern architecture hindered client experience, with multiple channels initiating various payment types that required complex processing.

Solution
Our team built a centralized payments hub to orchestrate data flows between payment initiation systems and product processors. The steps:
- Designed a flexible and scalable microservices-based architecture to facilitate translation, enrichment and processing of payment transactions
- Built a messaging layer to streamline data flows between systems, through support for various modes of interaction, e.g., MQ, API and file (canonical / industry standards such as NACHA, SWIFT, JSON, etc.)
- Introduced an API gateway to handle multiple payment types to enable channel agnostic payment capabilities
- Deployed a modular approach to support existing and new systems with isolation of core and product processors and avoid redundancies in capability builds
- Developed a React-based UI as the touchpoint for integrations between the payments hub and other systems

Outcomes
- A core payments engine capable of seamlessly integrating with multiple, complex systems
- Superior client experience, resulting from a holistic view spanning initiation, payment rails, and clearing
- A modernized payments platform that is ISO 20022-compliant and future-ready for processing and reporting needs
- Faster implementation of functionalities for payment processors

Our experts can help you find the right solutions to meet your needs.
Automated Scheduling Bots Boost Productivity by 50%



Client
Leading supply chain brokerage
Goal
Automate the manual, supply-chain scheduling process to improve staff productivity and customer satisfaction
Tools and Technologies
UI Path Orchestrator, UI Path Assistant, Microsoft Power BI, Office 365
Business Challenge
Performing crucial supply chain logistics, a provider’s operations team was struggling due to the high volume of scheduling appointments with shippers, receivers, and carriers, which involve back and forth emails, phone calls, or manual data entry into multiple Transport Management Systems (TMS).
These appointment-scheduling complexities vary based on the parties involved, from sending an email requesting appointment times to accessing a TMS and selecting what’s available as per their schedule.
Lacking proper analytics, sales representatives were unable to pinpoint peak appointment times, track cancellation rates, or discern customer preferences, often leading to shipment delays and incurred detention charges.

Solution
- Deployed multiple rule-based, automated workflows to pull information from incoming appointment requests (from emails, web forms, etc.) and automatically input it into the various TMS used to book pick-up and delivery appointments
- Developed a Power BI dashboard to visualize appointment trends, peak times, and cancellation rates, providing insights into customer behaviors, including frequent reschedules, preferred times, and typical lead times for booking appointments
- Delivered a reusable solution that could be leveraged for other business areas

Outcomes
- Bots operating 24/7 have led to over 15,000 monthly appointments being scheduled, resulting in a 50% reduction in manual scheduling hours
- The productivity of the operations team has improved by 50%, enabling staff to concentrate on high-value tasks rather than manual appointment-booking
- The increased accuracy in scheduled appointments has significantly decreased detention charges, thereby boosting overall customer satisfaction

Our experts can help you find the right solutions to meet your needs.
Automated POD improves turnaround time 95%



Client
Leading supply chain brokerage
Goal
Automate Proof of Delivery documentation process to increase efficiency and accuracy in data upload, validation and invoicing
Tools and Technologies
UI Path Orchestrator, UI Path Document Understanding, Microsoft Power BI, Oracle Transportation Management
Business Challenge
Proof of Delivery (POD) is a document that confirms an order has arrived at its destination and was successfully delivered before the invoice can be billed for payment.
Lack of an electronic POD system leads to inefficient, manual processing due to varied legal and contractual documentation requirements, resulting in longer billing cycles. Diverse formats and layouts from different carriers complicate data extraction from paper-based PODs.

Solution
- Developed a Document Processing Bot with UI Path AI Center, leveraging Document Understanding and Optical Character Recognition for managing various carrier documents
- Optimized data models for major carriers, focusing on the top five document types that represent 80% of the volume
- Implemented UI Path Action Center's "Human in the Loop" to handle exceptions and conducted 6-8 weeks of rigorous training on the Document Understanding model to ensure accuracy and meet confidence targets

Outcomes
- Achieved a 95% reduction in POD turnaround time, dropping from 48 hours to 2 hours, significantly boosting customer satisfaction
- Enhanced productivity by 87.5%, confirming receipt and condition of freight efficiently
- Reached 80% process accuracy, with continuous enhancement via automatic retraining
- Cut the billing cycle by 35%, allowing immediate use of data for customer invoicing

Our experts can help you find the right solutions to meet your needs.
Unified automation strategy enhances efficiency



Client
Leading payroll and HR solutions provider
Goal
Develop automation strategy and framework that accommodates growth and ensures efficiency
Tools and Technologies
Ansible, AWS, Dynatrace, Gremlin, Groovy, Jenkins, Keptn, KICS, Python, Terraform
Business Challenge
The SRE (Site Reliability Engineering) shared services team faced a diverse set of needs relating to automation of infrastructure and services provisioning, configuration, and deployment.
The team was encountering resource constraints, as limited in-house expertise in certain automation tools and technologies was causing delays in meeting critical automation requirements. They also needed to ensure system reliability and were challenged to scale automation solutions to accommodate increasing demands as operations grow.

Solution
- Development of a comprehensive automation strategy to align with objectives, encompassing Terraform, Ansible, Python, Groovy, and other relevant technologies in the AWS environment
- Leveraging our expertise to bridge the knowledge gap, provide training, and augment the client team in handling complex automation tasks
- Implementation of a chaos engineering framework using Gremlin, Dynatrace, Keptn, and EDA tools, to proactively identify weaknesses and enhance system resilience
- Creation of a scalable automation framework that accommodates growing needs and ensures long-term efficiency

Outcomes
- A unified automation strategy that streamlined processes, reduced manual effort, and enhanced overall efficiency by 30%
- The implementation of chaos engineering and self-healing practices, which increased reliability between 20% and 50%
- A reduction in manual interventions along with improved efficiency that will result in cost savings of 25% - 50%

Our experts can help you find the right solutions to meet your needs.
Home » Services » Automation » Page 2

The evolution of quality engineering
Quality engineering in software development empowers organizations to achieve heightened quality, scalability and resilience.


Quality Assurance (QA) has long been essential in software engineering, ensuring the development of products and applications with established standards and metrics. But QA has been reactive, focusing on defect detection through manual and automated testing. With the evolution of software development technology and methodologies, the limitations of traditional QA are evident. This perspective paper delves into the evolution of Quality Engineering (QE), which has transformed the approach to software quality. QE goes beyond QA and Test Automation to integrate quality practices throughout the Software Development Lifecycle (SDLC); it also addresses complexities in modern architectures such as microservices and cloud environments.
The journey from QA to QE is marked by several key milestones. Initially, software testing was a separate phase, conducted after development was complete. With the advent of Agile and DevOps methodologies, the need for continuous testing and early defect detection became apparent. This shift fostered the evolution of testing practices, embedding quality checks throughout the development cycle with the emerging adoption of cloud-native modern architecture, paving the way for what is now called Quality Engineering. Unlike the reactive nature of traditional QA and QA Automation, QE represents a proactive and integrated approach throughout the development lifecycle.
Organizations can significantly enhance product or application quality, optimize development workflows, and mitigate risks by addressing QE concerns at every phase of the SDLC. They could leverage structured QE approaches as mentioned above, and focus on a holistic view of quality in modern architecture.
Read our Perspective Paper for more insights on the evolution of quality engineering in software development.
Contact
Our experts can help you find the right solutions to meet your needs.
Get in touchQuality engineering optimizes a DLT platform



Client
A leading provider of financial services digitization solutions
Goal
Reliability assurance for a digital ledger technology (DLT) platform
Tools and Technologies
Kotlin, Java, Http Client, AWS, Azure, GCP, G42, OCP, AKS, EKS, Docker, Kubernetes, Helm Chart, Terraform
Business Challenge
A leader in Blockchain-based digital financial services required assurance for non-GUI (Graphic User Interface), Command Line Interface (CLI), microservices and Representational State Transfer (REST) APIs for a Digital Ledger Technology (DLT) platform, as well as platform reliability assurance on Azure, AWS services (EKS, AKS) to ensure availability, scalability, observability, monitoring and resilience (disaster recovery). It also wanted to identify capacity recommendations and any performance bottlenecks (whether impacting throughput or individual transaction latency) and required comprehensive automation coverage for older and newer product versions and management of frequent deliveries of multiple DLT product versions on a monthly basis.

Solution
- 130+ Dapps were developed and enhanced on the existing automation framework for terminal CLI and cluster utilities
- Quality engineering was streamlined with real-time dashboarding via Grafana and Prometheus
- Coverage for older and newer versions of the DLT platform was automated for smooth, frequent deliverables for confidence in releases
- The test case management tool, Xray, was implemented for transparent automation coverage
- Utilities were developed to execute a testing suite for AKS, EKS, local MAC/ Windows/ Linux cluster environments to run on a daily or as-needed basis

Outcomes
- Automation shortened release cycles from 1x/month to 1x/week; leads testing time was reduced by 80%
- Test automation coverage with 2,000 TCs was developed, with pass rate of 96% in daily runs
- Compatibility was created across AWS-EKS, Azure-AKS, Mac, Windows, Linux and local cluster
- Increased efficiency in deliverables was displayed, along with an annual $350K savings for TCMs
- An average throughput of 25 complete workflows per second was sustained
- Achieved a 95th percentile flow-completion time that should not exceed 10 seconds

Our experts can help you find the right solutions to meet your needs.
Home » Services » Automation » Page 2

Productionizing Generative AI Pilots
Get scalable solutions and unlock insights from information siloed across an enterprise by automating data extraction, streamlining workflows, and leveraging models.


Enterprises have vast amounts of unstructured information such as onboarding documents, contracts, financial statements, customer interaction records, confluence pages, etc., with valuable information siloed across formats and systems.
Generative AI is now starting to unlock new capabilities, with vector databases and Large Language Models (LLMs) tapping into unstructured information using natural language, enabling faster insight generation and decision-making. The advent of LLMs, exemplified by the publicly-available ChatGPT, has been a game-changer for information retrieval and contextual question answering. As LLMs evolve, they’re not just limited to text. They’re becoming multi-modal, capable of interpreting charts and images. With a large number of offerings, it is very easy to develop Proofs of Concept (PoCs) and pilot applications. However, to derive meaningful value, the PoCs and pilots need to be productionized and delivered in significant scale.
PoCs/pilots deal with only the tip of the iceberg. Productionizing needs to address a lot more that does not readily meet the eye. To scale extraction and indexing information, we need to establish a pipeline that, ideally, would be driven by events, new documents generated and available, possibly through an S3 document store and SQS (Simple Queue Service), to initiate parsing of documents for metadata, chunking, creating vector embedding and persisting metadata and vector embedding to suitable persistence stores. There is a need for logging and exception-handling, notification and automated retries when the pipeline encounters issues.
While developing pilot applications using Generative AI is easy, teams need to carefully work through a number of additional considerations to take these applications to production, scale the volume of documents and the user-base, and deliver full value. It would be easier to do this across multiple RAG (Retrieval-Augmented Generation) applications, utilizing conventional NLP (Natural Language Processing) and classification techniques to direct user requests to different RAG pipelines for different queries. Implementing the capabilities required around productionizing Generative AI applications using LLMs in a phased manner will ensure that value can be scaled as the overall solution architecture and infrastructure is enhanced.
Read our perspective paper for more insights on Productionizing Generative AI Pilots.
Contact
Our experts can help you find the right solutions to meet your needs.
Get in touchHome » Services » Automation » Page 2

How Gen AI Can Transform Software Engineering
Unlocking efficiency across the software development lifecycle, enabling faster delivery and higher quality outputs.


Generative AI has enormous potential for business use cases, and its application to software engineering is equally promising.
In our experience, development activities, including automated test and deployment scripts, account for only 30-50% of the time and effort spent across the software engineering lifecycle. Within that, only a fraction of the time and effort is spent in actual coding. Hence, to realize the true promise of Generative AI in software engineering, we need to look across the entire lifecycle.
A typical software engineering lifecycle involves a number of different personas (Product Owner, Business Analyst, Architect, Quality Assurance/ Tech Leads, Developer, Quality/ DevSecOps/ Platform Engineers), each using their own tools and producing a distinct set of artifacts. Integrating these different tools through a combination of Gen AI software engineering extensions and services will help streamline the flow of artifacts through the lifecycle, formalize the hand-off reviews, enable automated derivation of initial versions of related artifacts, etc.
As an art-of-the-possible exercise, we developed extensions (for VS Code IDE and Chrome Browser at this time) incorporating the above considerations. Our early experimentation suggests that Generative AI has the potential to enable more complete and consistent artifacts. This results in higher quality, productivity and agility, reducing churn and cycle time, across parts of the software engineering lifecycle that AI coding assistants do not currently address.
Complementary approaches to automate repetitive activities through smart templating, leveraging Generative AI and traditional artifact generation and completion techniques can help save time, let the team focus on higher-value activities and improve overall satisfaction. However, there are key considerations in order to do this at scale across many teams and team members. To enable teams to become high-performant, the Gen AI software engineering extensions and services need to provide capabilities around standardization and templatization of standard solution patterns (archetypes) and formalize the definition and automation of steps of doneness for each artifact type.
Read our perspective paper for more insights on How Gen AI Can Transform Software Engineering through streamlined processes, automated tasks, and augmented collaboration, bringing faster, higher-quality software delivery.
Contact
Our experts can help you find the right solutions to meet your needs.
Get in touchNC Tech Association Leadership Summit 2024

Iris Software will participate in the exclusive, attendance-capped, annual Leadership Summit hosted by the NC Tech Association, along with its board of directors and advisors, on August 7 & 8, 2024. Our representative, Senior Client Partner, Michel Abranches, will be among the executives gathering for the Summit, at the Pinehurst Resort in Pinehurst, NC, to network and discuss a variety of topics relevant to tech leaders and the projects and associates they manage.
The theme of this year’s summit is Adaptive Leadership. The event includes keynote addresses, executive workshops, and two panel discussions on ‘Why digital transformation is more about people than technology’ and ‘Building resilient tech teams: the power of emotional intelligence.’
As a technology provider to Fortune 500 and other leading global enterprises for more than 30 years, Iris is a trusted choice for leaders who want to realize the full potential of digital transformation. We deliver complex, mission-critical software engineering, application development, and advanced tech solutions that enhance business competitiveness and achieve key outcomes. Our agile, collaborative, right-sized teams and high-trust, high-performance, award-winning culture ensure clients enjoy top value and experience.
Contact Michel Abranches, based in our Charlotte, NC office, or visit www.irissoftware.com for details and success stories about our innovative approach and how we are leveraging the latest in AI / Gen AI / ML, Automation, Cloud, DevOps, Data Science, Enterprise Analytics, Integrations, and Quality Engineering.
Contact
Our experts can help you find the right solutions to meet your needs.
Get in touchIndustries
Company
