AutoScout24 is Europe’s main automotive market platform that connects patrons and sellers of latest and used vehicles, bikes, and industrial automobiles throughout a number of European international locations. Their long-term imaginative and prescient is to construct a Bot Manufacturing unit, a centralized framework for creating and deploying synthetic intelligence (AI) brokers that may carry out duties and make choices inside workflows, to considerably enhance operational effectivity throughout their group.
From disparate experiments to a standardized framework
As generative AI brokers (programs that may purpose, plan, and act) change into extra highly effective, the chance to enhance inner productiveness for AutoScout24 was clear. This led to numerous engineering groups experimenting with the know-how. As AI innovation accelerated throughout AutoScout24, they acknowledged a chance to pioneer a standardized method for AI improvement. Whereas AutoScout24 had efficiently experimented with varied instruments and frameworks on Amazon Net Providers (AWS), they envisioned making a unified, enterprise-grade framework that might allow quicker innovation. Their purpose was to ascertain a paved path that might make it simpler for groups throughout the group to construct safe, scalable, and maintainable AI brokers. The AutoScout24 AI Platform Engineering group partnered with the AWS Prototype and Cloud Engineering (PACE) group in a three-week AI bootcamp. The purpose was to maneuver from fragmented experiments to a coherent technique by making a reusable blueprint, a Bot Manufacturing unit, to standardize how future AI brokers are constructed and operated inside their firm.
The problem: figuring out a high-impact use case
To floor the Bot Manufacturing unit blueprint in a tangible enterprise case, the group focused a major operational value: inner developer help. The issue was well-defined. AutoScout24 AI Platform engineers have been spending as much as 30% of their time on repetitive duties like answering questions, granting entry to instruments, and finding documentation. This help tax decreased total productiveness. It diverted expert engineers from high-priority characteristic improvement and compelled different builders to attend for routine requests to be accomplished. An automatic help bot was a really perfect first use case as a result of it wanted to carry out two core agent features:
- Information retrieval: Answering “how-to” questions by looking inner documentation, a functionality often known as Retrieval Augmented Technology (RAG).
- Motion execution: Performing duties in different programs, corresponding to assigning a GitHub Copilot license, which requires safe API integration, or “instrument use.”
By constructing a bot that might do each, the group may validate the blueprint whereas delivering quick enterprise worth.
Architectural overview
On this put up, we discover the structure that AutoScout24 used to construct their standardized AI improvement framework, enabling speedy deployment of safe and scalable AI brokers.
The structure is designed with a easy, decoupled circulation to ensure the system is each resilient and easy to keep up. The diagram gives a simplified view targeted on the core generative-AI workflow. In a manufacturing surroundings, extra AWS companies corresponding to AWS Id and Entry Administration (IAM), Amazon CloudWatch, AWS X-Ray, AWS CloudTrail, AWS Net Software Firewall (WAF), and AWS Key Administration Service (KMS) might be built-in to reinforce safety, observability, and operational governance.
Right here is how a request flows by the system:
- Person interplay through Slack: A developer posts a message in a help channel, for instance, “@SupportBot, can I get a GitHub Copilot license?“
- Safe ingress through Amazon API Gateway & AWS Lambda: Slack sends the occasion to an Amazon API Gateway endpoint, which triggers an AWS Lambda operate. This operate performs a necessary safety test, verifying the request’s cryptographic signature to substantiate it’s authentically from Slack.
- Decoupling through Amazon Easy Queue Service (SQS): The verified request is positioned onto an Amazon SQS First-In, First-Out (FIFO) queue. This decouples the front-end from the agent, bettering resilience. Utilizing a FIFO queue with the message’s thread timestamp because the MessageGroupId makes positive that replies inside a single dialog are processed so as, which is essential for sustaining coherent conversations.
- Agent execution through Amazon Bedrock AgentCore: The SQS queue triggers a Lambda operate when messages arrive, which prompts the agent working within the AgentCore Runtime. AgentCore manages the operational duties, together with orchestrating calls to the muse mannequin and the agent’s instruments. The Orchestrator Agent’s logic, constructed with Strands Brokers, analyzes the consumer’s immediate and determines the proper specialised agent to invoke—both the Information Base Agent for a query or the GitHub Agent for an motion request.
An important implementation element is how the system leverages AgentCore’s full session isolation. To keep up conversational context, the system generates a novel, deterministic sessionId for every Slack thread by combining the channel ID and the thread’s timestamp. This sessionId is handed with each agent invocation inside that thread. Interactions in a thread share this similar sessionId, so the agent treats them as one steady dialog. In the meantime, interactions in different threads get completely different sessionIds, holding their contexts separate. In impact, every dialog runs in an remoted session: AgentCore spins up separate sources per sessionId, so context and state don’t leak between threads. In apply, because of this if a developer sends a number of messages in a single Slack thread, the agent remembers the sooner elements of that dialog. Every thread’s historical past is preserved mechanically by AgentCore.
This session administration technique can be very important for observability. Based mostly on a novel sessionId, the interplay could be traced utilizing AWS X-Ray, which provides perception into the circulation – from the Slack message arriving at API Gateway to the message being enqueued in SQS. It follows the orchestrator’s processing, the decision to the muse mannequin, subsequent instrument invocations (corresponding to a knowledge-base lookup or a GitHub API name), and at last the response again to Slack.
Metadata and timing assist point out the circulation of every step to know the place time is spent. If a step fails or is sluggish (for instance, a timeout on an exterior API name), X-Ray pinpoints which step brought about the difficulty. That is invaluable for diagnosing issues shortly and constructing confidence within the system’s habits.
The answer: A reusable blueprint powered by AWS
The Bot Manufacturing unit structure designed by the AutoScout24 and AWS groups is event-driven, serverless, and constructed on a basis of managed AWS companies. This method gives a resilient and scalable sample that may be tailored for brand new use instances.
The answer builds on Amazon Bedrock and its built-in capabilities:
- Amazon Bedrock gives entry to high-performing basis fashions (FMs), which act because the reasoning engine for the agent.
- Amazon Bedrock Information Bases permits the RAG functionality, permitting the agent to hook up with AutoScout24’s inner documentation and retrieve info to reply questions precisely.
- Amazon Bedrock AgentCore is a key part of the operational facet of the blueprint. It gives the absolutely managed, serverless runtime surroundings to deploy, function, and scale the brokers.
This answer gives a major benefit for AutoScout24. As an alternative of constructing foundational infrastructure for session administration, safety, and observability, they use AgentCore’s purpose-built companies. This enables the group to concentrate on the agent’s enterprise logic somewhat than the underlying infrastructure. AgentCore additionally gives built-in safety and isolation options. Every agent invocation runs in its personal remoted container, serving to to stop information leakage between periods. Brokers are assigned particular IAM roles to limit their AWS permissions (following the precept of least privilege). Credentials or tokens wanted by agent instruments (corresponding to a GitHub API key) are saved securely in AWS Secrets and techniques Supervisor and accessed at runtime. These options give the group a safe surroundings for working brokers with minimal customized infrastructure.
The agent itself was constructed utilizing the Strands Brokers SDK, an open-source framework that simplifies defining an agent’s logic, instruments, and habits in Python. This mixture proves efficient: Strands to construct the agent, and AgentCore to securely run it at scale. The group adopted a complicated “agents-as-tools” design sample, the place a central orchestrator Agent acts as the primary controller. This orchestrator doesn’t include the logic for each attainable job. As an alternative, it intelligently delegates requests to specialised, single-purpose brokers. For the help bot, this included a Information Base agent for dealing with informational queries and a GitHub agent for executing actions like assigning licenses. This modular design makes it simple to increase the system with new capabilities, corresponding to including a PR overview agent with out re-architecting your entire pipeline. Operating these brokers on Amazon Bedrock additional enhances flexibility, because the group can select from a broad vary of basis fashions. Extra highly effective fashions could be utilized to complicated reasoning duties, whereas lighter, cost-efficient fashions are well-suited for routine employee brokers corresponding to GitHub license requests or operational workflows. This capability to combine and match fashions permits Autoscout24 to steadiness value, efficiency, and accuracy throughout their agent structure.
Orchestrator agent: constructed with Strands SDK
Utilizing the Strands Brokers SDK helped the group to outline the orchestrator agent with concise, declarative code. The framework makes use of a model-driven method, the place the developer focuses on defining the agent’s directions and instruments, and the muse mannequin handles the reasoning and planning. The orchestrator agent could be expressed in only a few dozen strains of Python. The instance snippet beneath (simplified for readability, not meant for direct use) reveals how the agent is configured with a mannequin, a system immediate, and a listing of instruments (which on this structure characterize the specialised brokers):
One other instance is the GitHub Copilot license agent. It’s carried out as a Strands instrument operate. The next snippet reveals how the group outlined it utilizing the @instrument decorator. This operate creates a GitHubCopilotSeatAgent, passes the consumer’s request (a GitHub username) to it, and returns the end result:
Key advantages of this method embody clear separation of issues. The developer writes declarative code targeted on the agent’s goal. The complicated infrastructure logic, together with scaling, session administration, and safe execution, is dealt with by Amazon Bedrock AgentCore. This abstraction permits speedy improvement and allowed AutoScout24 to maneuver from prototype to manufacturing extra shortly. The instruments record successfully makes different brokers callable features, permitting the orchestrator to delegate duties while not having to know their inner implementation.
The impression: A validated blueprint for enterprise AI
The Bot Manufacturing unit challenge delivers outcomes that prolonged past the preliminary prototype. It creates quick enterprise worth and establishes a strategic basis for future AI innovation at AutoScout24.The important thing outcomes have been:
- A production-ready help bot: The group deployed a practical Slack bot that’s actively lowering the guide help load on the AutoScout24 AI Platform Engineering Crew, addressing the 30% of time beforehand spent on repetitive duties.
- A reusable Bot Manufacturing unit blueprint: The challenge produces a validated, reusable architectural sample. Now, groups at AutoScout24 can construct a brand new agent by beginning with this confirmed template (Slack -> API Gateway -> SQS -> AgentCore). This considerably accelerates innovation by permitting groups to concentrate on their distinctive enterprise logic, not on reinventing the infrastructure. This modular design additionally prepares them for extra superior multi-agent collaboration, probably utilizing requirements just like the Agent-to-Agent (A2A) protocol as their wants evolve.
- Enabling broader AI improvement: By abstracting away the infrastructure complexity, the Bot Manufacturing unit empowers extra folks to construct AI options. A website knowledgeable in safety or information analytics can now create a brand new instrument or specialised agent and “plug it in” to the manufacturing facility while not having to be an knowledgeable in distributed programs.
Conclusion: A brand new mannequin for enterprise brokers
AutoScout24’s partnership with AWS turned fragmented generative AI experiments right into a scalable, standardized framework. By adopting Amazon Bedrock AgentCore, the group moved their help bot from prototype to manufacturing, whereas specializing in their Bot Manufacturing unit imaginative and prescient. AgentCore manages session state and scaling, so engineers can concentrate on high-value enterprise logic as a substitute of infrastructure. The end result is greater than a help bot: it’s a reusable basis for constructing enterprise brokers. With AgentCore, AutoScout24 can transfer from prototype to manufacturing effectively, setting a mannequin for a way organizations can standardize generative AI improvement on AWS. To begin constructing enterprise brokers with Amazon Bedrock, discover the next sources:
Concerning the authors
Andrew Shved is a Senior AWS Prototyping Architect who leads groups and clients in constructing and transport Generative AI–pushed options, from early prototypes to manufacturing on AWS.
Muhammad Uzair Aslam is a tenured Technical Program Supervisor on the AWS Prototyping group, the place he works carefully with clients to speed up their cloud and AI journeys. He thrives on diving deep into technical particulars and turning complexity into impactful, value-driven options.
Arslan Mehboob is a Platform Engineer and AWS-certified options architect with deep experience in cloud infrastructure, scalable programs, and software program engineering. He presently builds resilient cloud platforms and is obsessed with AI and rising applied sciences.
Vadim Shiianov is a Information Scientist specializing in machine studying and AI-driven programs for real-world enterprise functions. He works on designing and deploying ML and Generative AI options that translate complicated information into measurable impression. He’s obsessed with rising applied sciences and constructing sensible, scalable programs round them.

