Acerto is a Financial company based in Brazil focused on providing debt negotiation services for general customers
Solution Area
Artificial Intelligence
Industry
Financial
Client Location
Brazil
The Challenge
The Business Solution
The Solution
To implement this business solution, Avenue Code’s development team collaborated closely with the client to establish an infrastructure within Google Cloud capable of processing more than 1,000 tickets daily via an API. To achieve this, the team employed Vertex AI and the Gemini-pro model as the core components of the inference pipeline. Additionally, a serverless environment was implemented, leveraging API Gateway and Load Balancer to ensure robust security measures.
1
Vertex AI and Gemini-pro Model:
– Vertex AI serves as the backbone of the inference pipeline, providing a comprehensive suite of machine learning tools and services.
– The Gemini-pro model, specifically designed for natural language processing tasks, plays a crucial role in classifying and summarizing ticket subjects based on their descriptions.
2
Serverless Environment:
– Serverless architecture was adopted to eliminate the need for managing and maintaining servers, enhancing scalability and cost-effectiveness.
– API Gateway facilitates secure and scalable handling of API requests, ensuring efficient communication between different components.
– Load Balancer distributes incoming traffic across multiple instances, promoting high availability and preventing single points of failure.
3
Integrated API for Zendesk Connection:
– A secure API was developed to establish a seamless connection with Zendesk.
– This API enables the exchange of data between the custom solution and Zendesk, ensuring real-time synchronization of ticket information.
4
Serverless Solution for Request Handling:
– A serverless solution was implemented to handle the high volume of requests efficiently.
– Serverless functions scale automatically based on demand, optimizing resource utilization and minimizing costs.
5
RAG Solution with Cloud SQL Database:
The Results
LLM Performance:
Cost Reduction:
Achieved a 70% reduction in operational costs by deploying our application on GCP’s serverless platform. This eliminated the need for costly on-premises infrastructure and allowed us to scale our application dynamically based on demand.
Data Processing:
Optimized more than 90% of the workload by utilizing GCP’s powerful data processing tools, such as BigQuery and Dataflow. This enabled us to handle large amounts of data efficiently and cost-effectively.