UofTHacks X Submission

Ollie AI

An AI assistant designed to detect, translate, and summarize large amounts of text from images, leveraging Cohere's Generate API for advanced text processing capabilities.

What tools it uses

React
Vite
Google Cloud Platform
Translate API
Vision API
Netlify
Cohere API

Links

Github Demo Devpost Figma 

How I built it

Ollie was designed and conceptualized on Figma, then coded using HTML/CSS, JavaScript and Cohere as an API. Our tech stack is simply React for the front end and node.js for the back end. To achieve our translation and summarization we are chaining a pipeline of API requests from different providers to get a seamless unique user experience. First, to get the text, we used either Google Cloud’s vision API for image-to-text detection or a textfield for direct text input. Next, we took that text and translated it to the desired language using Google Cloud’s translation API for text language translations. Finally, we utilized the Cohere AI Javascript SDK to request their generated API for text summarizations and then presented that text to the user in various levels of verbosity.

Figma Wireframe Demo