This project leverages Groq's API and LLAMA 3.2 to provide lightning-fast, cost-effective AI solutions for image analysis and audio transcription.
-
AI Image Analysis
- Upload any image and receive instant AI-powered insights
- Enhance your workflow with cutting-edge image analysis
-
Audio Transcription
- Upload audio files for quick and accurate transcription
- Powered by advanced language models for high-quality results
- AI Backend: Groq API
- Language Model: LLAMA 3.2
- Frontend: Next.js with React
- Super Fast: Utilizes Groq's high-performance API for rapid results
- Cost-Effective: LLAMA 3.2 provides state-of-the-art performance at a fraction of the cost
- Versatile: Handles both image analysis and audio transcription in one application
- Node.js (v14 or later)
- pnpm
- A Groq API key
-
Clone the repository:
git clone https://github.com/yourusername/ai-image-audio-analyzer.git cd ai-image-audio-analyzer
-
Install dependencies:
pnpm install
-
Copy the
.env.example
file to.env.local
and fill in your API keys:cp .env.example .env.local
-
Edit
.env.local
and add your Groq API key and other necessary credentials.
-
Start the development server:
pnpm dev
-
Open http://localhost:3000 in your browser.
GROQ_API_KEY
: Your Groq API keyGROQ_BASE_URL
: Groq API base URL (default: https://api.groq.com/openai/v1)MODEL_VISION
: The vision model to use (default: "llama-3.2-11b-vision-preview")NEXT_PUBLIC_VOICE_FAST_MODEL
: The fast voice model (default: "llama-3.2-3b-preview")NEXT_PUBLIC_SERVER_URL
: Your server URL (default: "http://localhost:3000")
Contributions are welcome! Please feel free to submit a Pull Request.
This project is licensed under the MIT License - see the LICENSE file for details.