๐ฏ This is a submission for the Google AI Studio Multimodal Challenge ๐
๐ฟ๐ Echo Location Project ๐ฆ โจ
๐ What I Built
I built ๐ Echo Location ๐, a web app that reframes the concept of an animal identifier into an interactive and purpose-driven conservation quest! ๐ฏ๐ The experience isn't just about answering "What animal is this?" ๐ค; it's about solving the deeper problem of the disconnect between humanity and the natural world! ๐๐ฑ
๐จ It's actually meant for mobile...but in this case it can also be seen as used on desktop, tablet...etc. ๐ฑ๐ป๐ฒ
๐ชโจ๐โญ๐ซโก๐ฅ๐ฅ๐๐๐โจ
๐ The Experience ๐
Echo Location transforms every user into an "Eco-Scout" ๐งญ๐ฟ on a mission! When a user uploads a photo ๐ธ, video ๐ฅ, or even an audio clip ๐ต of wildlife, they're not just getting a name. They're filing a "Sighting Report" ๐ that initiates an immersive learning journey! ๐โจ
๐ค TECH MAGIC: The app uses Gemini 2.5 Pro and Flash ๐ง โก to perform a deep multimodal analysis!
๐ฏ Generated Field Report Features:
๐ฆ The animal's story
๐จ Official conservation status
โ ๏ธ Primary threats
๐ Simulated geolocation based on its environment
๐ฎ Gamification Elements:
This journey is gamified through a "Ranger's Field Journal" ๐ where users:
- ๐ Level up
- ๐ Earn badges for completing ecosystem collections
- ๐ Unlock "Hope Spotlights"โreal stories of conservation success! ๐
๐ช MOST IMPORTANTLY: Echo Location bridges the digital-to-real-world gap by issuing actionable "Field Missions," like plastic cleanups ๐๏ธ or pollinator pledges ๐, turning learning into tangible, positive environmental action! ๐ช๐
๐ฌ Demo
๐ฑ Deployed Link: https://echo-location-849419792496.us-west1.run.app/
๐ธ Screenshots:
โจ๐โญ๐ซโก๐ฅ๐ฅ๐๐๐ช๐โจ
๐ง How I Used Google AI Studio
Google AI Studio was the central nervous system ๐งฌ for developing Echo Location! Gemini 2.5 Pro ๐ค is not just a feature; it is the brain ๐ง and the heart โค๏ธ of the entire application!
๐ฏ WORKFLOW MAGIC: My primary workflow revolved around extensive prompt engineering in the AI Studio! ๐ ๏ธโจ
I crafted a detailed system prompt that establishes the persona of "Gem," ๐ our AI Field Biologist! This prompt instructs Gemini to act as an enthusiastic guide ๐บ๏ธ and to structure all its responses in the format of our "Field Report." ๐
๐ The Process
๐ THE MAGIC FLOW:
๐ฑ User uploads media (image, video, or audio)
โ
๐ Backend sends to Gemini 2.5 Pro and Flash
โ
๐ง Prompt directs chain-of-thought analysis
โ
๐ Generates immersive Field Report!
๐ฏ Chain-of-Thought Analysis Breakdown:
๐ Identify: Identify the species, its scientific name, and its conservation status
๐ฟ Analyze: Analyze the surrounding environment, flora, and context to deduce a probable ecosystem and geolocation
๐ Narrate: Weave this data into an engaging, educational story in Gem's persona
โก Act: Based on the species and its threats, generate a relevant, actionable "Field Mission" for the user
๐ช PRO TIP: Google AI Studio was indispensable for rapidly testing and refining these complex prompts! Being able to quickly iterate on inputs and outputs and then deploy the model via Cloud Run made the entire development cycle seamless! ๐
๐จ Multimodal Features
The multimodal capabilities of Gemini 2.5 Pro are what elevate Echo Location from a simple app to an immersive experience! ๐ชโจ
๐ Holistic Scene and Video Understanding
๐ฏ This is the core input! The app doesn't just recognize an animal; it understands the scene! ๐ฌ When a user uploads a video of a bear ๐ป catching a salmon ๐, Gemini comprehends:
๐ฏ The action (hunting)
๐ค The interaction between species
๐ Context differences (e.g., a lion resting ๐ด vs. a lion on the prowl ๐)
๐ MAGIC MOMENT: This contextual understanding allows "Gem" to generate narratives that are incredibly rich and specific to the moment captured by the user, making every Field Report unique and insightful! ๐ซ
๐ต Eco-Acoustic Analysis
๐ช A standout feature is the ability to accept audio uploads! ๐๏ธ A user can record:
๐ฆ Bird calls in their backyard
๐ฆ The sound of insects at night
๐ค GEMINI MAGIC: Gemini analyzes this soundscape to identify potential species ("That distinct call belongs to a Northern Cardinal!" ๐ฆ) and describe the health of the local ecosystem! ๐ฟ
This turns the user's own environment into a subject of discovery and makes them feel like a true field biologist ๐ฌ using advanced tools! ๐ ๏ธ
๐ Context-Driven Content Generation
The app demonstrates a powerful multimodal feedback loop! ๐ช๏ธ The visual ๐๏ธ or audio ๐ input is the catalyst for all generated content!
๐ฏ Examples in Action:
๐ข Sea Turtle Image โ doesn't just trigger a story about turtles; it triggers the generation of the "Plastic Patrol Mission" ๐๏ธ
๐ Garden Bee Audio โ triggers the generation of the "Pollinator Pledge Mission" ๐ป
๐ช THE SECRET SAUCE: This ensures that the conservation message and the call-to-action are always directly relevant to the user's discovery, creating a powerful and persuasive user experience that drives the app's core mission forward! ๐๐
โจ๐โญ๐ซโก๐ฅ๐ฅ๐๐๐ช๐โจ
Made with โค๏ธ by @williamhenryking and @linfordlee14 ! ๐จโ๐ป๐จโ๐ป
๐ Thanks for reading our submission! ๐๐
Top comments (2)
It was amazing collaborating on this project! Learned a lot about multimodal AI and Google AI Studio along the way. Proud of what we achieved together!
Echo Location Project is such a creative application of AI! Projects like this show how much potential there is in multimodal AI for real-world solutions.