- Three 15-year-old students from Santa Clara, California—Akhil Nagori, Evann Sun, and Lucas Shengwen Yen—developed AI-powered smart glasses that convert text to speech in real time.
- The device, built for under $100, uses a Raspberry Pi, camera, and custom-trained neural network to help visually impaired students access reading materials.
- The prototype took five months to complete and features over 90% accuracy with a 13-second load time.
- The students won a $10,000 prize at the Thermo Fisher Scientific Junior Innovators Challenge, where Nagori also received the Thermo Fisher Scientific Leadership Award and Sun earned the Lemelson Foundation Award for Invention.
Quick Summary
Three California teenagers developed AI-powered smart glasses that convert text to speech for visually impaired students. The device costs less than $100 to build.
Akhil Nagori, Evann Sun, and Lucas Shengwen Yen spent five months creating the prototype. The glasses use a Raspberry Pi computer board, camera, battery, and speakers to capture images, extract text, and play audio through built-in speakers.
The students entered their project in the Thermo Fisher Scientific Junior Innovators Challenge and won a $10,000 prize. The competition selects only 300 national finalists from approximately 2,000 applicants.
The glasses achieve over 90% accuracy with an average load time of 13 seconds. The team trained their software using 800 images from school textbooks collected under three lighting conditions.
Nagori also received the Thermo Fisher Scientific Leadership Award, while Sun won the Lemelson Foundation Award for Invention. The students received a $5,000 grant to scale production and plan to distribute glasses throughout California.
The Innovation Behind the Glasses
Three 15-year-old students from Santa Clara created text-to-speech smart glasses powered by artificial intelligence. The wearable technology helps visually impaired students access reading materials from any format.
The prototype required five months of development and cost less than $100 to build. Akhil Nagori explained the motivation behind the project: "Our main goal was to create an easy, cost-efficient way to transcribe text from any format for visually impaired students."
The glasses operate by taking pictures of text, extracting the content, and converting it to audio played through tiny speakers built into the frames. The device uses a Raspberry Pi computer board, camera, battery, and speakers. It also includes a small on-and-off switch.
Performance metrics show the prototype achieves over 90% accuracy when translating text to speech. Lucas Shengwen Yen highlighted the importance of speed: "One of the most important aspects of our project is the load time. And that averaged around 13 seconds."
The inspiration came from a personal experience. Nagori traveled to India to visit family, including his great-uncle who is visually impaired and works as a cashier. He observed: "He has all these boxes filled with these braille receipts. He has to go through them line by line. When I saw that, I said, 'There's got to be an easier way that's not so tedious.'"
Our main goal was to create an easy, cost-efficient way to transcribe text from any format for visually impaired students— Akhil Nagori
Technical Development Process
The students faced three main challenges during development: hardware design, software programming, and data collection through testing.
Hardware Design
Evann Sun led the hardware design using Fusion 360 CAD software and a 3D printer to create custom frames. The team researched average glasses dimensions for middle to high school students to ensure proper fit. Sun explained: "Since we're trying to have all of the components on the glasses, we had to custom-design the areas for all of them."
Battery life was a critical consideration. Sun stated: "We want students to use this for, give or take, the entire school day. We are really concerned about the battery life, especially when we are using such a small battery."
Software Training
Nagori custom-trained a convolutional recurrent neural network (CRNN) using a dataset of 800 images. The training data came from school textbooks and other educational materials, featuring colorful images and various font styles.
The team collected images themselves under three lighting conditions: classroom lighting, low lighting, and outdoor lighting to ensure the model worked in different environments.
Testing and Refinement
The students conducted extensive testing to improve accuracy. Sun described the process: "After all our hardware and software were done, we tested our software part. We would input images that we saw online or that we took ourselves into our software model. Then, it extracted the text and gave us an MP3 file, which we could use to improve our accuracy."
Competition Success and Challenges
The Thermo Fisher Scientific Junior Innovators Challenge represents one of the top STEM research competitions for students. The selection process is highly competitive: middle school students must first compete at local science or engineering fairs, where judges nominate the top 10% of projects. Approximately 2,000 winners apply for the national competition, but only 300 are selected. From those, judges choose 30 finalists who present their research in Washington, D.C.
The team faced significant obstacles during their journey. Nagori admitted: "We had a lot of all-nighters." The most dramatic challenge occurred hours before their presentation when the glasses suffered a critical failure.
Yen described the crisis: "On the flight there, some of our soldering came off the Raspberry Pi. Without the soldering, nothing worked, and the glasses wouldn't start up. We were all in panic mode."
The team scrambled to fix the device. Yen continued: "The night before we were presenting, my dad ran to the closest mechanic store and got a soldering iron. The three of us put on masks, hunched over, and fixed it."
Despite not receiving an initial nomination, judges from the state level attended their competition and recognized the project's value. Sun reflected on the experience: "I think it really taught us that even if we don't get what we want the first time, as long as we work hard and stay committed, we can come back and be better."
In addition to the $10,000 prize for the glasses project, Nagori earned the Thermo Fisher Scientific Leadership Award, and Sun received $10,000 for the Lemelson Foundation Award for Invention.
Future Plans and Scaling
The research remains in the prototype phase, but the students have ambitious expansion plans. They received a $5,000 grant to scale their glasses and reach more of their community.
Nagori outlined their immediate goals: "We're currently working on implementing a lot of our glasses throughout California."
The team is already preparing for larger-scale production. Nagori noted they have "a big 3D printer in my garage right now with 30 Raspbe[rry Pis], 30 cameras, 30 batteries."
The project demonstrates how innovative technology can be developed affordably, contrasting with Silicon Valley companies that are raising billions to develop wearable AI products. While major tech firms invest massive resources, these teenagers proved that impactful solutions can be created for under $100.
Their work addresses a real-world problem for visually impaired students, providing a cost-effective alternative to expensive assistive technology. The combination of AI-powered text recognition, real-time audio conversion, and portable design creates a practical tool for educational accessibility.
"One of the most important aspects of our project is the load time. And that averaged around 13 seconds"
— Lucas Shengwen Yen
"He has all these boxes filled with these braille receipts. He has to go through them line by line. When I saw that, I said, 'There's got to be an easier way that's not so tedious'"
— Akhil Nagori
"We found the average glasses dimensions for middle to high school students. Since we're trying to have all of the components on the glasses, we had to custom-design the areas for all of them"
— Evann Sun
"We want students to use this for, give or take, the entire school day. We are really concerned about the battery life, especially when we are using such a small battery"
— Evann Sun
"We had to take those images ourselves in three different lighting conditions that mimicked classroom, low lighting, and outdoor lighting"
— Akhil Nagori
"After all our hardware and software were done, we tested our software part. We would input images that we saw online or that we took ourselves into our software model. Then, it extracted the text and gave us an MP3 file, which we could use to improve our accuracy"
— Evann Sun
"We had a lot of all-nighters"
— Akhil Nagori
"On the flight there, some of our soldering came off the Raspberry Pi. Without the soldering, nothing worked, and the glasses wouldn't start up. We were all in panic mode"
— Lucas Shengwen Yen
"The night before we were presenting, my dad ran to the closest mechanic store and got a soldering iron. The three of us put on masks, hunched over, and fixed it"
— Lucas Shengwen Yen
"I think it really taught us that even if we don't get what we want the first time, as long as we work hard and stay committed, we can come back and be better"
— Evann Sun
"We're currently working on implementing a lot of our glasses throughout California"
— Akhil Nagori
Frequently Asked Questions
How much did it cost to build the AI smart glasses?
The three teenagers built their AI-powered text-to-speech smart glasses for under $100 using a Raspberry Pi computer board, camera, battery, speakers, and 3D-printed frames.
What awards did the students win for their invention?
The team won a $10,000 prize at the Thermo Fisher Scientific Junior Innovators Challenge. Additionally, Akhil Nagori received the Thermo Fisher Scientific Leadership Award, and Evann Sun won the Lemelson Foundation Award for Invention, which also came with $10,000.
How do the glasses work?
The glasses use a camera to take pictures of text, which is processed by a custom-trained AI neural network. The extracted text is converted to audio and played through built-in speakers. The device achieves over 90% accuracy with an average load time of 13 seconds.

