AI-powered assistant converting your speech to forms using Google Cloud Run & Functions
Figma
Maze
3 Weeks
Developers
Designers
Build a creative, enjoyable, and valuable solution for optimizing a resource-intensive business. Use Cloud Run, Cloud Functions, and Workflows.
We built a serverless web application, reimagining the workflow of multi-input forms.
Teaming up with 3 developers (Santiago Uriarte, Brandon Huynh, Daniel Vu) and 1 designer (Tiffany Trinh), we joined the Google Cloud Easy as Pie Serverless Hackathon, aiming to foster skill growth and collaborative innovation together.
Objective: Build and deploy serverless apps with Cloud Run, Cloud Functions, and Workflows.
Prompt: "What’s your most sticky, time-consuming, budget-draining business function? Imagine that hot mess transformed into a sleek, simple process."
Our team brainstormed together to decide which business problem we wanted to tackle. We decided to address the pain points of forms using speech-to-text to:
Help increase workflow efficiency
Address the documentation needs of customers
Create a scalable SaaS product applicable to several industries
We surveyed 30 participants about their recent experiences with forms and speech-to-text. Our research goals were to identify pain points with forms. We were also curious about the value of speech-to-text and applicable markets.
What is the current experience like with forms?
How often do users use speech-to-text?
Where do users typically use forms?
say filling out forms are tedious
rarely or never use speech recognition
say speech recognition is convenient
Repetitive
Writing Burdens
Time-Consuming
Government
Healthcare
Education
Although 67% of participants rarely or never use speech-to-text, 53% found it convenient. This reveals an opportunity to expand speech recognition usage.
Due to time constraints, we needed to focus on high-priority features. We crossed out lower priority features, making our main features of the app to be:
Logging in as an existing user
Creating a SMART form
Accessing the transcribed form
Creating a template
Editing a form
Downloading a completed form
Using Figma, we created low-fidelity wireframes to depict potential designs for the app. We aimed the designs to be:
Familiar and learnable since this was not an ordinary transcription app.
Optimized for mobile device to allow for flexible usage wherever, whenever.
Simple for ease of use and navigation when performing the MVP tasks.
Using Maze, we conducted 4 remote, moderated and 10 remote, unmoderated usability tests. Our goals were to assess user navigation and ease.
Select a Template
Create a SMART Form
View SMART Form
Success Rate
Misclick Rate
Duration
"Nothing really confused me, it was pretty straightforward."
"I thought it was a list, but I wasn’t sure if I had to click anywhere else before...it seemed like an unclickable list."
"Anything where you purchase at a counter...or anything if they want to take note what the customer or buyer needs, this would be useful."
"...The way it was laid out and when it gave me the time accessed and when the file was opened, it kind of confused me because I wasn’t sure if this was a template or a file that I made."
Based on feedback, we realized that users needed more context and clarification. Refinements done to address these main concerns are as shown:
Intended to be a scalable SaaS product, the branding is modern, clean, and professional.
I enjoyed building a demo product with this team. This experience taught me how to work with others in an agile environment. A few thoughts:
The app usability improved by 10% based on task success & duration, test exits, and misclicks.
If I had more time:
I would have explored an onboarding flow to guide users on what the app is about and its main features.
Explored motion design and animation for the audio waves to increase accessibility.
What I learned:
Attending weekly stand-ups and reviews helped us identify roadblocks, expectations, and new learnings.
Bouncing off ideas with one another during brainstorming sessions helped spur creative solutions.