Research

We investigated existing solutions, relevant technologies, and theoretical frameworks to inform the project’s direction.

Technology Review

For solutions to a system like Blueprint AI, we considered two aspects.

1. Presence of Coding

This aspect refers to how much the user will need to code to get their frontend designs out and working, also taking into account a user’s programming skill level. The increasing prevalence of low-code/no-code products such as Wix and Microsoft PowerApps signals a shift in traditional software development workflow, reducing the barriers to entry for software development. This is greatly empowering for the average user as the technical skills needed are vastly reduced [11]. Furthermore, the estimated low-code/no-code market size was USD 6.78 billion in 2022, and is projected to grow significantly in the next decade [12].

While a low-code/no-code design choice would make sense given the current market, this would limit the flexibility and control a user has when designing their frontend. The user would be limited to pre-designed components and structures, which is restrictive [13]. Additionally, employing a low-code/no-code design would not make our student-developed project standout against industry standard solutions like Wix and PowerApps which are already established.

On the other hand, incorporating too many elements of coding required by the user would cater towards the more experienced developers and would be more time-consuming; a counter-intuitive design choice for Blueprint AI.

With the intended users being those who require a quick, reliable way to scaffold and generate frontend designs - regardless of their know-how and technical expertise in frontend development, our approach would need to cater to this type of user. As such, an approach requiring more coding than conventional low-code/no-code platforms but at the same time not be a necessary component for the frontend generation. This comes in the form of the exporting feature where users can edit the exported code in the frontend framework they need, a welcomed feature over low-code/no-code platforms like Wix which lack an exporting feature. Users of all programming backgrounds would thus find that using Blueprint AI is valuable in terms of the time savings it affords and also the flexibility in editing the frontend code generated. Figure 5 depicts a general indication of where Blueprint AI falls in the spectrum of coding required by the user.

Diagram 1

Figure 5. Coding spectrum approach used for Blueprint AI

2. Prominence of AI

Currently, popular industry solutions like Wix, Visily and PowerApps do feature AI tools to help generate frontend designs (as discussed under Section 1.1), but these AI tools are relegated to being an additional feature rather than being integrated throughout every aspect that the user interacts with when designing. We aim to get Blueprint AI to stand out from existing solutions through making AI more prominent; studies have shown that AI is extremely useful in expediting frontend development, thereby alleviating workloads through automation [14].

In order to make AI more prominent, from the start, users are prompted to tell the AI what they want through text, image or both. We also plan to include a chat bot for the user to interact with the AI by telling it what changes to make. This allows for prompts by the user that can be tailored to the users specific needs, enhancing the flexibility and control that the user has as well - beyond manually tweaking the frontend source code themselves. Additionally, the AI also recommends different pages that the users need as frontend design requires multiple pages, on average 12-30 pages for a full-fledged website - according to [15].

With this, at every step, AI is integrated seamlessly into the frontend development experience, streamlining the development process. With AI-driven UI generation, users are afforded efficiency and flexibility. Figure 6 depicts the continuous integration of AI in Blueprint AI.

Diagram 2

Figure 6. Continuous integration of AI in Blueprint AI

One of the standout features of v1.0 was its integration as a VSCode extension. The sheer user base VSCode has, amounting to over 2.5 million active monthly users in 2017 [16] makes it an attractive choice to deploy Blueprint AI as a VSCode extension. The free-to-use IDE hosts over 4000 extensions [16] and has comprehensive documentation on how to deploy our project as an extension [6]. Thus, having Blueprint AI as a VSCode extension means it runs on any device that supports VSCode, removing hardware dependency concerns and allowing our project to reach a wide range of developers.

For UI layout recognition, we needed software that could effectively and efficiently extract text description from given screenshots of websites. This text description would then be used to describe to the AI what kind of website the user wants to build. With this as a priority, we considered the following approaches:

1. Traditional Image Processing (OpenCV, cv2)

OpenCV (Open Source Computer Vision Library) is a computer vision and machine learning software library. It is widely used for real-time image and video processing, and it provides a comprehensive set of tools for tasks like object detection, facial recognition, motion tracking, and more. Its cross platform capabilities and common programming language interfaces (C, C++, Python, Java) make it easier to develop with [17].

Pros Cons
No licensing cost,
Extensive Documentation,
Versatile as it supports a wide range of computer vision tasks.
Steep learning curve owing to wide range of advanced tools available

For Blueprint AI, there is simply no need for real-time image processing. This allowed us to narrow the search for libraries that related to text description extraction from an image, overcoming the need to pore through advanced libraries for heavy-duty image processing. Given that OpenCV is also widely used in industry, it made OpenCV a suitable choice.

Deep learning based Optical Character Recognition (OCR):

OCR is a technology that converts images (e.g., scanned documents, photos, or screenshots) into text that computers can read and edit . While OCRs can defer in the types of images they can extract from, the main function of an OCR is to allow computers to recognise and extract text from images. Some OCRs leverage AI and deep learning in the image processing and text detection and recognition front, allowing for more robust and impressive functionality [18]. This use of AI by OCR is inline with Blueprint AI’s central theme and abstract, as such, we researched on OCRs that specialised in extracting from screenshots that have deep learning capabilities.

We narrowed our options down to two competing solutions: EasyOCR and Tesseract. Both are open-source systems which use AI for processing. The table below shows what each OCR can do.

OCR Description Standout Features Drawbacks
EasyOCR Python based OCR which has support for over 80 languages.
Built on Pytorch and easy to implement and use [19].
Faster on GPU due to its deep learning-based architecture.
Provides higher accuracy for noisy, complicated images. This can be useful for complicated website layouts such as Chinese e-commerce platforms, which are known to be more intricate compared to Western counterparts [20].

Handles multilingual text in a single image effectively
Limited fine-tuning options.
Tesseract PA traditional OCR engine developed in part by Google with comprehensive documentation found in [21].

Optimized for CPU and generally faster on CPU.
Highly customisable, supports custom training and fine tuning.

Affords high accuracy with clean input.
Struggles with noisy or distorted text.

Requires separate language data files for each language.

We decided to use EasyOCR, with the primary reason being the ease of installation and implementation with Python and its simple API. Tesseract would require more set-up to get it running and implementation would be a bit more tedious. Additionally, while both OCRs would need some form of image pre-processing to be done before character recognition, EasyOCR requires less pre-processing. Lastly, while AI capabilities are an important factor for Blueprint AI, fine-tuning the OCR to produce the highest quality result is not a priority as we want our system to focus more on efficiency and speed for our users to rapidly scaffold their frontend development.

In Blueprint AI, we use cv2 (OpenCV) to clean up images before running them through EasyOCR for text extraction. Think of cv2 as a tool that prepares the image—adjusting brightness, removing noise, and sharpening text—so it’s easier to read. This step is important because raw screenshots can be cluttered or blurry, making it harder for the OCR to recognize text accurately. Once the image is prepped, EasyOCR, a deep learning-based tool, quickly scans and extracts the text, even from complex website layouts. This setup helps us efficiently turn website screenshots into clear, structured descriptions that the AI can use to generate front-end designs.

1. Programming Languages and Frameworks

We decided to continue the use of React, CraftJS and FluentUI from v1.0. React’s component-based architecture allows us to break down UI components into smaller, reusable pieces. This modularity is perfect for generating dynamic UIs, as components can be composed and reused across the application, ensuring ease of generating UIs dynamically based on user and AI input.

CraftJS provides a drag-and-drop interface for building and editing UIs. This makes it easier for non-technical users or designers to create and modify layouts without writing code. Also, CraftJS integrates with React components which makes the system more cohesive.

With these frameworks, we used Typescript instead of Javascript because the static typing of Typescript allowed for catching errors at compile time rather than runtime, easing the development and debugging process significantly. This also helps ensure the project remains scalable and maintainable as it grows as Typescript is better for organisation and refactoring.

2. AI API

For Blueprint AI’s backend, we leverage LangChain and Axios to streamline AI interactions and data fetching. LangChain is a powerful framework designed to work with large language models (LLMs), making it easier to build AI-driven applications by managing prompt chaining, memory, and model selection. This allows Blueprint AI to create more dynamic and context-aware interactions when generating front-end designs. Meanwhile, Axios, a popular JavaScript library for making HTTP requests, ensures smooth and efficient communication with APIs, including OpenAI’s models. Its promise-based structure simplifies data handling, making API calls more reliable and easier to manage. Together, LangChain and Axios help optimize AI integration, ensuring a seamless and responsive user experience.

We have also decided to stick with using OpenAI for the AI API. OpenAI contains many different models suitable for this project, such as DALLE, a generative AI model that creates images from text descriptions, allowing for image generation. OpenAI also boasts industry leading models that are highly advanced and intelligent, outperforming other competitors like Meta, Amazon and Microsoft [22]. With the impressive line-up of AI models, users also have more flexibility in managing their own tokens to manage their costs. Combined with the fact that OpenAI is also highly popular and widely used [23], and integrating OpenAI API calls into the backend is relatively straightforward, Blueprint AI will use OpenAI for its AI integration.

Summary of Technical Decisions

Category Final Choice Reason
Programming Frameworks React (CraftJS, FluentUI, MaterialUI) Best for component-based UI generation
Programming Language Typescript Easier to catch errors, debug, refactor and extend
Backend AI Integration OpenAI, LangChain, Axios Best Natural Language Processing capabilities
Programming Frameworks React (CraftJS, FluentUI, MaterialUI) Best for component-based UI generation
OCR/Text Recognition EasyOCR + cv2 Best for text extraction from screenshots.
Deployment Platform VSCode Developers work within VSCode, improving workflow, eliminating hardware dependencies

References

[1] K. Goldstein, “Wix ADI: How Design AI Elevates Website Creation for Everyone,” Wix Blog, Jun. 19, 2022. https://www.wix.com/blog/wix-artificial-design-intelligence

[2] “Power Pages – Website Builder | Microsoft Power Platform,” Microsoft.com, 2024. https://www.microsoft.com/en-us/power-platform/products/power-pages#tabs-pill-bar-ocb9d4_tab4 (accessed Mar. 05, 2025).

[3] “Edit code with Visual Studio Code for the Web (preview),” Microsoft.com, Mar. 07, 2024. https://learn.microsoft.com/en-us/power-pages/configure/visual-studio-code-editor?WT.mc_id=powerportals_inproduct_portalstudio2 (accessed Mar. 05, 2025).

[4] I. Bouchrika, “Mobile vs Desktop Usage Statistics for 2020/2021,” Research.com, Feb. 24, 2021. https://research.com/software/mobile-vs-desktop-usage

[5] lukejlatham, “GitHub - lukejlatham/vscode-rapid-gui,” GitHub, 2024. https://github.com/lukejlatham/vscode-rapid-gui/tree/main (accessed Mar. 15, 2025).

[6] “Webview API,” code.visualstudio.com. https://code.visualstudio.com/api/extension-guides/webview

[7] Karl-Bridge-Microsoft, “Windows UI Library (WinUI) - Windows apps,” learn.microsoft.com. https://learn.microsoft.com/en-us/windows/apps/winui/

[8] Y. Zhang, Y. Yuan, and A. C.-C. Yao, “Meta Prompting for AI Systems,” arXiv.org, 2023. https://arxiv.org/abs/2311.11482

[9] A. Mendonça, “Monolithic Modular Architecture: Modular Folder Organization,” Medium, Jul. 22, 2023. https://medium.com/@abel.ncm/monolithic-modular-architecture-modular-folder-organization-4cbf97175ab4 (accessed Mar. 18, 2025).

[10] S. Watts, “The Importance of SOLID Design Principles,” BMC Blogs, Jun. 15, 2020. https://www.bmc.com/blogs/solid-design-principles/

[11] A. Saxena, “How Policy Reforms And Interoperability Are Reshaping Value-Based Care,” Forbes, Mar. 20, 2025. https://www.forbes.com/councils/forbestechcouncil/2025/03/20/how-policy-reforms-and-interoperability-are-reshaping-value-based-care/ (accessed Mar. 20, 2025).

[12] “Low-code Development Platform Market Size Report, 2030,” www.grandviewresearch.com. https://www.grandviewresearch.com/industry-analysis/low-code-development-platform-market-report

[13] “What’s Wrong with Low and No Code Platforms?,” www.pandium.com. https://www.pandium.com/blogs/whats-wrong-with-low-and-no-code-platforms

[14] Manoj Kumar Dobbala, “Rise of Generative AI: Impacts on Frontend Development,” Sep. 09, 2023. https://www.researchgate.net/publication/382639784_Rise_of_Generative_AI_Impacts_on_Frontend_Development

[15] J. Bracamontes, “How Many Pages Should A Website Have?,” Acumenstudio.com, Feb. 06, 2023. https://acumenstudio.com/how-many-pages-should-a-website-have/

[16] S. McBreen, “Visual Studio Code 2.6M Users, 12 months of momentum and more to come.,” Nov. 15, 2017 code.visualstudio.com. https://code.visualstudio.com/blogs/2017/11/16/connect

[17] OpenCV, “About OpenCV,” OpenCV, 2018. https://opencv.org/about/

[18] Amazon, “What is OCR (Optical Character Recognition)? - AWS,” Amazon Web Services, Inc. https://aws.amazon.com/what-is/ocr/

[19] A. Mahajan, “EasyOCR: A Comprehensive Guide,” Medium, Oct. 29, 2023. https://medium.com/@adityamahajan.work/easyocr-a-comprehensive-guide-5ff1cb850168

[20] J. Nielsen and Y. Cheng, “Are Chinese Websites Too Complex?,” Nielsen Norman Group, Nov. 06, 2016. https://www.nngroup.com/articles/china-website-complexity/ (accessed Mar. 22, 2025).

[21] tesseract-ocr, “tesseract-ocr/tesseract,” GitHub, Oct. 20, 2019. https://github.com/tesseract-ocr/tesseract

[22] “OpenAI - Quality, Performance & Price Analysis | Artificial Analysis,” Artificialanalysis.ai, 2025. https://artificialanalysis.ai/providers/openai

[23] E. H. Schwartz, “OpenAI confirms 400 million weekly ChatGPT users - here’s 5 great ways to use the world’s most popular AI chatbot,” TechRadar, Feb. 22, 2025. https://www.techradar.com/computing/artificial-intelligence/openai-confirms-400-million-weekly-chatgpt-users-heres-5-great-ways-to-use-the-worlds-most-popular-ai-chatbot (accessed Mar. 22, 2025).

This research forms the foundation for Blueprint AI’s development, ensuring we use the best technologies to optimize AI-assisted front-end generation.