Document Type

Honors Thesis

Abstract

MobileCLIP is a compact model that connects images and text, enabling it to perform tasks like image identification and question answering without needing to be retrained for each new task. Although it’s designed to be lightweight, it still runs slowly on regular computers without a powerful graphics card (GPU). This research focuses on making MobileCLIP run faster and smaller by using post-training quantization, which reduces the model’s precision after training without hurting performance. We combined several strategies: analyzing which parts of the model are more sensitive to changes, applying targeted adjustments to its structure, and running everything using CPU-only tools. Our tests showed we could shrink the entire model by 53% (from 397MB down to 186MB) and cut its overall processing time by 5%, while keeping its accuracy nearly the same. This work shows that advanced AI models can run on everyday computers, opening the door to more accessible, energy-efficient AI.

Disciplines

Computer and Systems Architecture

Publication Date

Fall 2025

Faculty Mentor of Honors Project

Marnim Galib

License

Creative Commons Attribution 4.0 International License
This work is licensed under a Creative Commons Attribution 4.0 International License.

Share

COinS
 
 

To view the content in your browser, please download Adobe Reader or, alternately,
you may Download the file to your hard drive.

NOTE: The latest versions of Adobe Reader do not support viewing PDF files within Firefox on Mac OS and if you are using a modern (Intel) Mac, there is no official plugin for viewing PDF files within the browser window.