- Language: Python 3.10+
- AI Framework: LlamaIndex (RAG Orchestration)
- LLM Engine: GPT4All (Local Model Execution)
- Model: Llama 3.2 1B-Instruct (Quantized GGUF)
- Embeddings: HuggingFace
bge-small-en-v1.5 - Interface: Streamlit
- API: Adzuna Job Search API
- Secure Upload: Drag and drop your
.pdfor.docxresume safely. - Instant Analysis: Get a professional breakdown of your resume's strengths and weaknesses.
- Rating System: Receive scores on content quality, keywords, and market relevance.
- Live Job Matching: Automatically search for active job vacancies that match your profile.
- Data Privacy: Experience AI without an internet connection (except for job searching).
R: Rerun the application.C: Clear the cache (useful if you change the model file).Enter: Confirm file upload or trigger the analysis button when focused.
The development followed a "Privacy-First" philosophy.
- Environment Setup: Created a virtual environment to manage complex AI dependencies.
- Model Selection: Tested different GGUF models to find the perfect balance between speed and reasoning (Llama 3.2 1B).
- Bridge Building: Developed a
CustomLLMadapter class to connect the GPT4All local engine to the LlamaIndex framework. - RAG Implementation: Configured a vector store to index resume data for precise context retrieval.
- API Integration: Connected the Adzuna service to bridge the gap between "AI analysis" and "Real-world action."
- LLM Adapters: How to write custom classes to make incompatible libraries work together.
- Resource Management: Using
@st.cache_resourceto keep large models in RAM, preventing system lag. - Prompt Engineering: How to guide small 1B models to stay focused on specific tasks without "hallucinating."
- RAG Architecture: The power of giving an AI a specific "memory" (the resume) rather than relying on its general knowledge.
- Cover Letter Generation: Add a feature to draft a personalized cover letter for the found jobs.
- Multi-language Support: Expand the prompt templates to analyze resumes in Portuguese and Spanish.
- Database Integration: Allow users to save their analysis history locally using SQLite.
- Better Models: Testing Llama 3.2 3B or Mistral models for even deeper reasoning if hardware allows.
- Clone & Enter:
git clone https://github.com/JoyceAcacioPedro/My-Career.git
- Install Requirements:
pip install streamlit llama-index gpt4all requests docx2txt llama-index-embeddings-huggingface
- Model Path:
Open
My_Career.pyand ensure themodel_pathmatches your local GPT4All folder. - Run:
streamlit run My_Career.py
"Watch how My Career Advisor transforms a raw resume into a targeted career strategy in seconds."
Developed with 🧠 by Joyce Pedro