Quickstart
Start building LLM-powered applications in under 5 minutes
Getting Started Steps
Create your account
Visit the Scoopika platform and create a free account.
Install Scoopika in your project
For React or NextJS applications, refer to this guide.
First, install the main SDK:
This SDK is built for server-side usage and to deploy your AI agents as API endpoints, you’ll learn how to do that later, don’t worry it’s a simple copy-paste process that We’ll walk through together.
We also have packages for client-side usage and React applications that We’ll discover later (requires a deployed AI agent).
Generate access token
You need to generate your own access token from the platform by clicking on the settings icon in the top-left corner.
Once your access token is generated you should copy it and add it to your environment variables like this:
This access token will be used to authenticate your application with Scoopika only when sending audio or url inputs to your AI agents, or enabling voice responses, otherwise it won’t be used.
Audio files, URLs, and voice responses generation happens on our servers because it requires running some heavy AI models.
Initialize Scoopika
Now you can initialize Scoopika in your app, you only need one line of code:
Connect LLM provider
In order to run AI agents, You need to connect your LLM provider using your own API key in order to use the LLMs provided by it.
Scoopika currently supports a number of LLM providers that’s growing everyday, You can connect multiple LLM providers in your app and use multiple models provided by different providers, you can see how to connect these providers here. and here are a few examples:
You’ll see a number of supported LLMs under each provider in the platform, but you can really use any model provided by these providers, you just need to pass its name when creating your AI agents, so let’s do it!
Create AI agents
Don’t know what an AI agent is? see this page.
AI agents are just LLMs with access to extra tools and functionalities, in Scoopika AI agents are the main component that can be used for text, voice, and object generation.
Create an AI agent is simple:
You can also pass extra optional properties to your Scoopika instance and AI agent to give it access to long-term memory or knowledge stores (hosted services we provide to make your development much easier):
If you want to give your AI agents to have long-term memory and save conversations history, you can create a memory store here.
If you want to extend the model’s knowledge with information from files, PDFs, or websites, you can create a knowledge store here, upload your info, and just add its ID.
Let’s see a full usage example with Groq:
But you’re here for the heavy stuff right? now in the above code you’ll get built-in error recovery without any additional code, but what about multimodal inputs and streaming responses? well, it’s just the same:
But what about conversations history? once you connect a memory store to your AI agents, you can pass sessions IDs when running an AI agent, learn more here.
You can pass and control a lot of things using the run options other than the session ID:
Deploy AI agent
Now we have an AI agent, let’s deploy it as an API endpoint so we can use it from the client-side with built-in streaming support and data validation!
The process of deploying an AI agent might change a little bit based on the framework you’re using, but it should work with any web framework that supports HTTP streaming. for example, when using Express this is how you do it:
See how to deploy your AI agent based on your framework here.
What’s Next?
Once you have a deployed AI agent, you can start using the client-side library to run it, the process is so simple and just the same as running the AI agent on the server-side, learn more here.
Learn more about:
Deploying AI agents and using them on the client-side.
Text generation and building AI assistants.
Object generation and data extraction.
Was this page helpful?