Gen AI Content Generation Model Using LLM Without Internet
Dr. B. Ravikrishna, Vinuthna Taduru, Arvind Reddy Ravula, V Shiva Kumar, G Samuel Ashish
The Gen AI content generation model uses the LLAMA-3.1-8B model where 8B implies 8 billion
parameters to effectively produce customized content. It generates the content based on three inputs:
topic, word count, and complexity level, which can be easy, medium, or hard. The front-end development
is done by using Streamlit and CSS-based interface to make the model user-friendly. Backend functions
are supported by LangChain and CTransformers modules for proper communication. It is totally an
offline-working model that minimizes reliance on servers or even internet access in favor of increased
security. This pre-training will have been extensive over datasets that try to uncover all patterns while
the fine-tuning is toward applications in various tasks, thereby perfecting text summarization, question
answering, content generation, or even classification tasks.