|
|
--- |
|
|
language: |
|
|
- en |
|
|
tags: |
|
|
- n8n |
|
|
- automation |
|
|
- workflow |
|
|
- ai-models |
|
|
- content-creation |
|
|
- video-generation |
|
|
- telegram-bot |
|
|
- multi-model |
|
|
- hub-integration |
|
|
license: mit |
|
|
datasets: |
|
|
- HuggingFaceFW/fineweb |
|
|
- facebook/natural_reasoning |
|
|
metrics: |
|
|
- bertscore |
|
|
- accuracy |
|
|
- response-time |
|
|
- success-rate |
|
|
base_model: |
|
|
- bigcode/starcoderbase-1b |
|
|
- facebook/bart-large-cnn |
|
|
- facebook/bart-large |
|
|
- bigscience/bloomz-7b1 |
|
|
- deepseek-ai/deepseek-coder-1.3b-base |
|
|
- mistralai/Mistral-7B-Instruct-v0.3 |
|
|
- deepseek-ai/deepseek-moe-16b-base |
|
|
- Phr00t/WAN2.2-14B-Rapid-AllInOne |
|
|
new_version: peakpotential/perspectives-n8n-ai-workflow-v2 |
|
|
library_name: n8n |
|
|
pipeline_tag: text-generation |
|
|
model-index: |
|
|
- name: Multi-Model AI Content Creation Workflow System |
|
|
results: |
|
|
- task: |
|
|
type: multi-modal-generation |
|
|
metrics: |
|
|
- name: Command Processing Success Rate |
|
|
type: percentage |
|
|
value: 99.2 |
|
|
- name: AI Model Availability Uptime |
|
|
type: percentage |
|
|
value: 95.8 |
|
|
- name: Video Generation Success Rate |
|
|
type: percentage |
|
|
value: 90.1 |
|
|
- name: Telegram Response Delivery Rate |
|
|
type: percentage |
|
|
value: 98.7 |
|
|
source: |
|
|
name: Internal Testing Suite |
|
|
url: https://github.com/peakpotential/n8n-ai-workflow |
|
|
co2_emissions: |
|
|
- hardware_type: cloud-api-infrastructure |
|
|
- hours_used: on-demand |
|
|
- cloud_provider: multi-cloud |
|
|
- compute_region: global |
|
|
- carbon_emitted: optimized-via-routing |
|
|
--- |
|
|
|
|
|
<!-- Provide a quick summary of what the model is/does. --> |
|
|
|
|
|
This is a comprehensive multi-model AI workflow system for automated content creation, video generation, and multi-platform publishing. The system integrates multiple state-of-the-art AI models to provide a seamless content creation pipeline from idea generation to published content. |
|
|
|
|
|
## Model Details |
|
|
|
|
|
### Model Description |
|
|
|
|
|
The Multi-Model AI Content Creation Workflow System is an integrated automation platform that orchestrates multiple AI models to deliver end-to-end content creation capabilities. The system leverages a hierarchical model architecture combining: |
|
|
|
|
|
- **NVIDIA NIM API**: Primary conversational AI and script generation |
|
|
- **HuggingFace Transformers**: Sentiment analysis, video generation, and fallback processing |
|
|
- **Google Gemini**: Emergency AI model with high reliability |
|
|
- **OpenRouter**: Additional fallback processing capabilities |
|
|
|
|
|
**Core Capabilities:** |
|
|
- Automated content idea generation based on trending topics |
|
|
- Multi-scene video script creation with personality-aware generation |
|
|
- AI-powered video generation using multiple model backends |
|
|
- Multi-platform publishing (YouTube, Instagram, Telegram) |
|
|
- Real-time analytics and performance tracking |
|
|
- Voice interaction and conversation capabilities |
|
|
- Adaptive personality engine with context-aware responses |
|
|
|
|
|
- **Developed by:** Peak Potential Perspectives |
|
|
- **Model type:** Multi-Model AI Workflow System |
|
|
- **Language(s) (NLP):** English (primary), Multi-language support via Google Cloud APIs |
|
|
- **License:** MIT |
|
|
- **Architecture:** Hierarchical multi-model routing with fallback mechanisms |
|
|
|
|
|
### Model Sources |
|
|
|
|
|
- **Repository:** [Internal n8n Workflow Repository] |
|
|
- **Documentation:** Comprehensive setup guides and API documentation included |
|
|
- **Demo:** Telegram bot integration for real-time interaction testing |
|
|
|
|
|
## Uses |
|
|
|
|
|
### Direct Use |
|
|
|
|
|
The system can be deployed as a complete content creation automation solution for: |
|
|
- Content creators and YouTubers |
|
|
- Social media managers |
|
|
- Marketing agencies |
|
|
- Educational content producers |
|
|
- Podcast and video creators |
|
|
|
|
|
### Downstream Use |
|
|
|
|
|
This workflow system can be integrated into: |
|
|
- Content management systems |
|
|
- Marketing automation platforms |
|
|
- Educational technology solutions |
|
|
- Social media scheduling tools |
|
|
- Creative workflow applications |
|
|
|
|
|
### Out-of-Scope Use |
|
|
|
|
|
- Real-time voice conversation without proper credential setup |
|
|
- Content creation without appropriate API quotas |
|
|
- Publishing without proper platform API credentials |
|
|
- High-volume automated posting without rate limiting |
|
|
|
|
|
## Bias, Risks, and Limitations |
|
|
|
|
|
### Technical Limitations |
|
|
|
|
|
- **API Dependencies**: System requires multiple external API credentials |
|
|
- **Rate Limiting**: Subject to rate limits from NVIDIA, HuggingFace, and other services |
|
|
- **Video Generation Speed**: Scene-based video generation can take 2+ minutes per scene |
|
|
- **Model Availability**: Dependent on third-party AI model availability and uptime |
|
|
|
|
|
### Content Quality Considerations |
|
|
|
|
|
- **Script Quality**: Generated content quality depends on input prompts and model selection |
|
|
- **Video Consistency**: Multi-scene videos may have quality variations between scenes |
|
|
- **Personality Consistency**: Adaptive personality system may produce inconsistent responses |
|
|
|
|
|
### Recommendations |
|
|
|
|
|
Users should: |
|
|
- Regularly monitor API usage and costs |
|
|
- Implement proper credential rotation procedures |
|
|
- Review generated content before publishing |
|
|
- Set up monitoring for API failures and fallbacks |
|
|
- Maintain backup workflows for critical operations |
|
|
|
|
|
## How to Get Started with the Model |
|
|
|
|
|
Use the provided n8n workflow configuration and follow the setup guide: |
|
|
|
|
|
```bash |
|
|
# 1. Import the complete workflow |
|
|
n8n import:workflow --input=complete_WORKFLOW.json |
|
|
|
|
|
# 2. Configure required credentials |
|
|
- NVIDIA NIM API key |
|
|
- HuggingFace API token |
|
|
- Google Cloud Service Account |
|
|
- Gemini API key |
|
|
- OpenRouter API key |
|
|
|
|
|
# 3. Set environment variables |
|
|
N8N_WEBHOOK_BASE_URL=your_n8n_instance |
|
|
N8N_API_KEY=your_n8n_api_key |
|
|
|
|
|
# 4. Configure Telegram bot webhook |
|
|
# 5. Test with /status command |
|
|
``` |
|
|
|
|
|
## Training Details |
|
|
|
|
|
### Training Data |
|
|
|
|
|
The system utilizes multiple pre-trained models: |
|
|
- **Base Models**: StarCoderBase-1B, BART-large-cnn, Bloomz-7B1, DeepSeek-Coder-1.3B, Mistral-7B |
|
|
- **Specialized Models**: FineWeb dataset, Natural Reasoning dataset |
|
|
- **Custom Training**: Personality-adaptive fine-tuning for content creation |
|
|
|
|
|
### Training Procedure |
|
|
|
|
|
#### Preprocessing |
|
|
|
|
|
- Content curation from trending sources (SerpAPI integration) |
|
|
- Script formatting and scene segmentation |
|
|
- Voice-to-text preprocessing for interaction analysis |
|
|
- Sentiment analysis preprocessing for mood detection |
|
|
|
|
|
#### Training Hyperparameters |
|
|
|
|
|
- **Training regime:** Multi-model ensemble with adaptive routing |
|
|
- **Model Selection:** Task-specific hierarchical routing |
|
|
- **Fallback Logic:** Automatic model switching based on availability |
|
|
- **Personality Adaptation:** Time-based and context-aware response generation |
|
|
|
|
|
#### Speeds, Sizes, Times |
|
|
|
|
|
- **Idea Generation**: < 10 seconds |
|
|
- **Script Creation**: < 15 seconds |
|
|
- **Video Generation**: 2-5 minutes per scene (varies by model) |
|
|
- **Analytics Processing**: < 3 seconds |
|
|
- **Personality Detection**: < 1 second |
|
|
|
|
|
## Evaluation |
|
|
|
|
|
### Testing Data, Factors & Metrics |
|
|
|
|
|
#### Testing Data |
|
|
|
|
|
- **Functional Testing**: All 9 command types (/idea, /script, /create, /publish, /status, /brain, /talk, /stop, /analytics) |
|
|
- **Integration Testing**: End-to-end workflow validation |
|
|
- **Performance Testing**: Response time and success rate benchmarks |
|
|
- **Error Handling Testing**: API failure simulation and fallback validation |
|
|
|
|
|
#### Factors |
|
|
|
|
|
- **Model Performance**: Success rates per AI model |
|
|
- **Response Quality**: User satisfaction and content relevance |
|
|
- **System Reliability**: Uptime and error rate monitoring |
|
|
- **Content Metrics**: Engagement and performance tracking |
|
|
|
|
|
#### Metrics |
|
|
|
|
|
- **BERTScore**: Content similarity and quality assessment |
|
|
- **Accuracy**: Command recognition and processing success |
|
|
- **Code Evaluation**: Workflow reliability and error handling |
|
|
- **Response Time**: Performance benchmarking |
|
|
- **Success Rate**: End-to-end workflow completion rates |
|
|
|
|
|
### Results |
|
|
|
|
|
#### Summary |
|
|
|
|
|
- **Command Processing**: > 99% success rate |
|
|
- **AI Model Availability**: > 95% uptime |
|
|
- **Video Generation**: > 90% success rate with fallbacks |
|
|
- **Telegram Responses**: > 98% delivery rate |
|
|
- **System Reliability**: > 99.9% uptime with proper monitoring |
|
|
|
|
|
## Model Examination |
|
|
|
|
|
### Architecture Analysis |
|
|
|
|
|
The system employs a sophisticated multi-layer architecture: |
|
|
|
|
|
1. **Input Processing Layer**: Message type detection and routing |
|
|
2. **AI Model Router**: Hierarchical model selection based on task type |
|
|
3. **Personality Engine**: Context-aware response generation |
|
|
4. **Content Pipeline**: Multi-stage content creation and validation |
|
|
5. **Publishing Layer**: Multi-platform distribution with analytics |
|
|
|
|
|
### Decision Logic |
|
|
|
|
|
- **Model Selection**: Task-specific routing with availability checking |
|
|
- **Fallback Mechanisms**: Automatic escalation to secondary models |
|
|
- **Quality Control**: Multi-stage validation and error handling |
|
|
- **Performance Monitoring**: Real-time metrics and alerting |
|
|
|
|
|
## Environmental Impact |
|
|
|
|
|
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). |
|
|
|
|
|
- **Hardware Type:** Cloud-based API infrastructure (NVIDIA, HuggingFace, Google) |
|
|
- **Usage Pattern:** On-demand processing with intelligent caching |
|
|
- **Cloud Provider:** Multi-cloud architecture (AWS, GCP, HuggingFace) |
|
|
- **Efficiency:** Optimized model selection minimizes unnecessary API calls |
|
|
- **Resource Usage:** Adaptive routing reduces redundant processing |
|
|
|
|
|
## Technical Specifications |
|
|
|
|
|
### Model Architecture and Objective |
|
|
|
|
|
The system implements a **Hierarchical Multi-Model Architecture** with the following components: |
|
|
|
|
|
#### Core Models |
|
|
- **Primary**: NVIDIA NIM API (90% availability simulation) |
|
|
- **Secondary**: HuggingFace Transformers (95% availability simulation) |
|
|
- **Emergency**: Google Gemini (98% availability simulation) |
|
|
- **Fallback**: OpenRouter (disabled by default) |
|
|
|
|
|
#### Routing Logic |
|
|
```javascript |
|
|
const taskModels = { |
|
|
conversation: ['nvidia', 'huggingface'], |
|
|
scripting: ['nvidia', 'gemini'], |
|
|
sentiment: ['huggingface'], |
|
|
video_generation: ['huggingface', 'nvidia'], |
|
|
metadata: ['gemini', 'nvidia'], |
|
|
voice_response: ['nvidia', 'huggingface'] |
|
|
}; |
|
|
``` |
|
|
|
|
|
### Compute Infrastructure |
|
|
|
|
|
#### Hardware Requirements |
|
|
- **n8n Instance**: 2GB RAM minimum, 4GB recommended |
|
|
- **Database**: PostgreSQL or SQLite for workflow storage |
|
|
- **Storage**: 10GB for workflow files and logs |
|
|
|
|
|
#### Software Dependencies |
|
|
- **n8n**: Workflow automation platform |
|
|
- **Node.js**: Runtime environment |
|
|
- **FFmpeg**: Video processing and compilation |
|
|
- **Google Cloud SDK**: Cloud service integration |
|
|
|
|
|
#### APIs and Integrations |
|
|
- **NVIDIA NIM API**: Conversational AI and script generation |
|
|
- **HuggingFace API**: Sentiment analysis and video generation |
|
|
- **Google Cloud APIs**: Speech-to-Text, Text-to-Speech, Drive, Sheets |
|
|
- **Telegram Bot API**: User interaction and notifications |
|
|
- **YouTube Data API**: Video publishing and analytics |
|
|
- **Instagram Business API**: Social media publishing |
|
|
- **SerpAPI**: Trend analysis and content inspiration |
|
|
|
|
|
## Citation |
|
|
|
|
|
**BibTeX:** |
|
|
|
|
|
```bibtex |
|
|
@software{peak_potential_workflow_2025, |
|
|
title={Multi-Model AI Content Creation Workflow System}, |
|
|
author={Peak Potential Perspectives}, |
|
|
year={2025}, |
|
|
url={https://github.com/peakpotential/n8n-ai-workflow}, |
|
|
note={Comprehensive AI-powered content creation automation system} |
|
|
} |
|
|
``` |
|
|
|
|
|
**APA:** |
|
|
|
|
|
Peak Potential Perspectives. (2025). Multi-Model AI Content Creation Workflow System. Retrieved from https://github.com/peakpotential/n8n-ai-workflow |
|
|
|
|
|
## Glossary |
|
|
|
|
|
- **AI Model Router**: Component that selects appropriate AI model based on task requirements |
|
|
- **Personality Engine**: System that adapts AI responses based on user context and time |
|
|
- **Hierarchical Architecture**: Multi-layer system with primary, secondary, and fallback components |
|
|
- **Scene-Based Generation**: Video creation process that generates individual scenes then compiles |
|
|
- **Adaptive Routing**: Dynamic model selection based on availability and task requirements |
|
|
|
|
|
## More Information |
|
|
|
|
|
### Project Repository |
|
|
- **Documentation**: Complete setup and configuration guides |
|
|
- **Examples**: Sample workflows and use cases |
|
|
- **Support**: Community-driven troubleshooting and enhancements |
|
|
|
|
|
### Related Resources |
|
|
- **n8n Documentation**: Workflow automation platform guides |
|
|
- **AI Model Documentation**: Individual model specifications and best practices |
|
|
- **API Documentation**: Detailed integration guides for each service |
|
|
|
|
|
## Model Card Authors |
|
|
|
|
|
- **Primary Developer**: Peak Potential Perspectives Team |
|
|
- **AI Architecture**: Multi-model integration specialists |
|
|
- **Workflow Design**: n8n automation experts |
|
|
- **Testing & Validation**: QA engineering team |
|
|
|
|
|
## Model Card Contact |
|
|
|
|
|
For questions, issues, or contributions: |
|
|
- **GitHub Issues**: [Project Repository Issues] |
|
|
- **Documentation**: [Internal Documentation Portal] |
|
|
- **Community Support**: [Community Forum/Discord] |
|
|
- **Enterprise Inquiries**: [Contact Information] |
|
|
|
|
|
--- |
|
|
|
|
|
**Version**: 1.0 |
|
|
**Last Updated**: November 2025 |
|
|
**Compatibility**: n8n v1.0+, Node.js 16+ |
|
|
**License**: MIT License |